This content was uploaded by our users and we assume good faith they have the permission to share this book. If you own the copyright to this book and it is wrongfully on our website, we offer a simple DMCA procedure to remove your content from our site. Start by pressing the button below!
0 will obey the requirement: the function of the scalar argument t iP(t) = tp(A' + ei) if*, i.e. aggregation is not optimal. The lemma is valid. If the strict inequality z > z is fulfilled, then aggregation with the vector a is not optimal. This completes theorem (29.4). §29.2. Let us formulate the algorithm for solving problem (29.1). The process of "improving" the aggregating vector with the functional decreas ing for the disaggregated solutions admissible to (29.1) is based on formula (29.17) and relation (29.19). Step 0: Choose a' 0 ' the vector of aggregation weights and the parameter /3 e [0,1]. Let 1 = 0, £° = o o . Step 1: Solve aggregative problem (29.3) for the weight vector a'''. Let {xo,Xi,yi}, (i = 1, k) be a solution and : ft' -> ft i.e. i-e. y = 2/0 with x = x®, y = 2/? with x — x® from the con dition for achieving the vector criterion minimum Vo(xo,x\,yo,yi,ip) = {Vj{x0, %x, j/o, 2/i, v),---,V0N(x0, Xi, 2/o, Vi, 2}JV{arg, x?, 0, / i , . . . ,/w; ¥>} G Mi n M2 implies that V 0 s (xo,£i, 2/o.J/i-V0) = Vy(*o.*l»3toifiiV)It might be shown that the three definitions of the Pareto-efficient el ement {XQ,X1, J/§> y?' V°} G Mi n M 2 are equivalent, and sometimes other names are used: the non-dominant, unimprovable, Pareto-optimal or min imal element. Definition 42.3: The element {x%,x%,y%,y^, f i > ° 0)) V o ( x o , i i , » o , » i , v ) << VoVio.J/o.afi.yi.V 3/?i i/5°) such that for all s = 1 , . . . , N the inequalities ) ,!(t^) M\V\Mi M\C\M2 L s = l,...,/V *lif8.flr^} li >V^xlxlylylv V VJixlxlylyl?)1) C,(V ),), 0,yl^)>C s(V 00 ) 0, m 0, m 2 (x°) = min k(x°)- - min k(x), k{x°) k(x), lex1 igx 1 '- 1 where <j>s{x) = £
Dantzig- Wulf Decomposition
73 73
must monotonously decrease with 0 < t < 0 . The technique of selecting the number of 0 is denoted by E. Each pair of the techniques T, E determines a block programming method. On the basis of these methods we can also formulate combined procedures for block programming when in the process of solving a problem, different techniques T, E are employed. §10.4. By way of example we consider the ^ ( A J minimization procedure based on the employment of the generalized gradient lowering. Thus, we have problem (10.1) - (10.4). Assume that the set S = {x\A2x = b2, x > 0} is bounded. If this is not the case, it is always possible to add to system (10.3) the inequality of the form g Xi < M and select M so large that even one of the optimal solutions'of problem (10.1) - (10.4) will satisfy this inequality. Form the Lagrange function £(A,x), A = ( A i , . . . ,Ami) for problem (10.1), (10.2) and consider the following problem: n
minmax£(A,:r) = minmax A>0 x€S
A>0 xeS
mi
/
n
\
+ y AA. [ bb\ - V o La ax * ] ,(10.27) (10.27) V c *c *x* + 1 J t]
Y * > Yt— J' (\ ) - Y h* ^
^—' Li=l .1=1
J= J = ll
\V
i=l i=l
j // ..
,
wherre&},\,. . 6 ^ ) = bt; ;jy, ; = T^ ; = V ^ are the elements of the matrix Ai. Study the function separately n n
£*(A) £*(A) = = max max l€
Tn\ mi
//
n n
V V
^ clxl + +£ ^ AXJ3 (ffeb)i~ - Y ^ aa^Xi £ **,• iJXi) J
L i=i
i=l
V
t=i
(10.28) (10.28)
/.
Evidently, the function £*(A) is the function ip(A) dessribed above. Inas much as the polyhedron S is bounded, the function £*(A) is denned for any A. Owing to applying maximum operation to the convex set of linear forms the function £*(A) is convex. And since S is a polyhedron, £*(A) is a piecewise linear functiono FoF A > 0 and x d x where x = (xu.x\ .xn) is the feasible solution of problem (10.1) - (10.4) n
mi
X
/
b
n
a
\
Xi
> ■ * JY^cc*Ax>■
£*(A) > C(A,X) £(A,x) ==£<**< £<**<++^YA jAI )6] --Y^ a^j ^ i=\
j=l
V
i=l 1=1
n n
/
i=l
Thus, £*(A) is bounded below for A > 0, and since £*(A) is piecewise linear, there exists A* such that £*(A*) = min£*(A). A>0
74
Systems
Optimization
Methodology
Consider the problem dual to problem (10.1) - (10.4)
( TTl\
TTl\
i=\
mi
m.2
\
(10.29)
Y^bjXi 7Tl2 + ^blak), Y.bUi+Y,^). i=i
fc=i
(10.29) /
(10.30)
7Tl2
k=\
Y^a^K + ^al^^Cj,
j = UT,
(10.30)
(10.31) X{ i =k=\ l , m i ; crfc > 0, k — l,m2i=\ > 0, A,>0, i=T~^; ak>0, k = l~^. (10.31) L e m m a : An optimal solution to problem (10.29) - (10.31) of the form (A%S*) = ( A * , . . . , A ; i , a * , . . . , < 2 ) (A*,S*)
corresponds to any optimal solution of the problem (10.27). Proof.41 From this lemma it follows that the solution of problem (10.29) - (10.31) reduces to minimization of the function £*(A) with the constraints A > 0. UX*{K) = (irJ(A),... ,x;(A)) is the vector on which a maximum in (10.28) is achieved, then n
m\i m
/
n
\
ti=l =l
= \l j=
V
i=l
/
1 £*(A) ^ AAJ J 6^ a k * * (AA>) ) Cix*(A) + £*(A) = = ^5>**(A) +£ -- ] £I>X(
(10.32)
(10.32)
From expression (10.32) it follows that the generalized gradient lowering direction at the point A coincides with the direction opposite to that of the vector ff
(10.33)
(10.33)
Formulate an algorithm for solving the problem (10.29) - (10.31) dual with respect to the original problem (10.1) - (10.4). Step 1: Let the iteration number / = 1, fix the value A ^ , in particular, AW can be taken as equal to 0. Step 2: Solve problem (10.28), obtain the vector X*{A«) and the vector of an optimal solution to the dual problem £(A<')). n
Step 3: Calculate the vector a, {(b] - £ o L x J ( A « ) } , 1=1
Step j . Derive A<;+1) from the formula
A +
" "="{ A '"- t 'i?ii}'
; = U^.
Dantzig-Wulf
Decomposition
75
where h > 0 is the value of the step along the lowering direction, p is the operator leaving nonnegative coordinates without any change and transforming negative coordinates into 0. Go to step 2. CO
If kt is selected so that h — 0, £ k, = oo, then, in this case, '->°°
i=i
converging £(A*'>) to an optimal value of problem (10.27) is guaranteed: (;) £*(A ) —>£*(A*). C*(A{1) /—t-OO /—•■oo
Here the pair of vectors ( A « , S ( A ^ ) ) converges to an optimal solution of problem (10.29) - (10.31). Terminate. §11. T R A N S P O R T A T I O N PROBLEM SOLUTION B Y T H E DECOMPOSITION METHOD §11.1. One of the most interesting special problems of linear program ming is the transportation problem, the classical formulation of which takes the form: 71
771
^5 3^ 5c3 ^ci3-Xinj n~* i nm ',n ' j= l
(11.1)
i=l
n
^xlJ=al,
i=T^L,
(11.2)
3=1 771
1 n &*« 3' J =- M 5£ 3 x*a« ==& i> ' ,'
(11.3) (11.3)
1=1
Xij>0, «y>0,
i = T~fH\ l,m;
j = XH. l,n.
(11.4)
We consider the transportation problem (11.1) - (11.4) as the Block pro gramming problem in which conditions are partitioned into two blocks. In the notation in §9, conditions (11.2) correspond to the submatrix Au and conditions (11.3) to the submatrix A2. The decomposition method was used by Williams 8 to solve the transportation problems of large dimension ality. Following the scheme of the decomposition method, consider the set 5 prescribed by conditions (11.3) - (11.4). This set represents a polyhe dron with the nodes that are obtained if for each j one of x{j is assumed to
76
Systems Optimization
Methodology
be equal to bh the rest being equal to null. Denote the quantity of nodes N = m" by k kfessl.JV. = TJf. j), *x k = = {x\ {*%}, An arbitrary point of the polyhedron S is represented as: N TV
(11.5)
x = £zfcx*, fc=l fc=i
N N
Y*zh=l,
zZ k>0, k > 0
l,N. k = Tjf.
,
(11.6)
k=l
Substitute expression (11.5) into conditions (11.1) and (11.2). Obtain the coordinating or Z-problem: N
^fkZk
-* min,
Jt=i
N 2jPfc2fc = a, zz** > 00,,
(11.7)
kfe= = l,iV, TJt,
where a = (oi,aa,... n
,am)T,
T7X
fc = 1,JV j=l i = l
and the p* vector components are determined from the equality n
^fc == 1l,JV, ^-
Pl = YlxiJ'
= Vm. l,m. ii =
(11.8) (11.8)
J=I
w The convexity condition £ ^ = 1 for the transptrtation problem is the consequence of equalities in (11.7). Indeed, summing equalities in (11.7) by rows we get m
N
m
X) £pt^ =t =!]««•• i = i k=i i i=l k=l
t=l
Dantzig-
Wulf Decomposition
77
Employing (11.8), we have N N
m m nn
m m
fc=i i i=i fc=l = l j j=i =l
1=1 i=\
and since x% satisfies conditions (11.3) and (11.4), we derive the equality N
m
k=l k=l
1=1 i=l
a zk
m
J2{J2 *) = 12a* i=l i=l
from which the condition for convexity follows. The A A -problem is constructed according to the scheme of the decom position method. The linear form appears as: nn
m m
CA(x) = ^^ ^£( t^ej i - \i)Xij CA(X) Xi)xij, ,
(11.9) (11.9)
jJ = l 2 = 1
where A = ( A 1 ; . . . ,A m ) is a vector of dual variables. The A^-problem definition domain is determined by conditions (11.3) and (11.4). From the goal function inseparability (11.9) and independence of con straints (11.3) it follows that the X A -problem falls into n independent subproblems of the form: m
^(cij ^(cg
- Xi)xij ~* min,
2=1 m
-
' / , xv ~ "j"i >
/ , Xij 2=1
(11.10) (11.10)
=
Xij > 0, x^ 0.
The complete optimal solution of the AVproblem constitutes a vector in which for each j only one component is different from null and equal to br The number S of the nonzero component is calculated for each j from the relation CSJ - As = min (eg CSJ (eg - A,-). t1=1,771 =l,7n
The vector p*. of the Z-problem conditions corresponds to the solution xk* of the AVproblem. The estimate of this vector with respect to the current
78
Systems
Optimization
Methodology
basis is equal to the value of linear form (11.9) calculated on the optimal solution of the AVproblem: n
m
A t . =5353(c tj -A < )*&
(11.11)
j = l »=i j=i i=l
If A*. = 0, then the Z-problem basis under study is optimal, and an optimal solution to the original transportation problem is expressed in the form fc = l
where z*k is the optimal solution of the Z-problem. If A*. < 0, the iter ation process should be continued, here the vector pk. with the minimal characteristic difference Afe. is introduced into the Z-problem basis. The pfc. vector components are calculated by the formula n
#.=£*&,
i == i ^\,m. r.
(11.12) (n.12)
3=1 3=1
The algorithm described on a basis of the decomposition method has proved to be particularly effective for transportation problems under condition n
The complete optimal solution of the AVproblem is given by (( 00 ,, .. .. .. ,,b\,--b l , . . . ,, 00 ,, 00 ,, .. .. .. ,, SS ,, .. .. -. ,, 00 ,, .. .. .. ,, 00 ,, .. .. .. ,&«,.., ,&«,.., ,, 00' )) .. « ' i v ' * m m « ' m" ' * ' v v v v
m
v
m
m
Dantzig- Wulf Decomposition
79
Step 2: From formula (11.11) calculate the estimate A*, with respect to the current basis of the vector pk. corresponding to the solution of the X A -problem. Step 3: If Afe- > 0, then the current basic program z*k of the Z-problem is optimal. Go to step 5. Step 4: If Afc. < 0, then we introduce into the Z-problem basis the vector pfc. the components of which are calculated with formula (11.12). The column obtained is multiplied by the inverse basic matrix, and a new inverse matrix and a vector of dual variables A are determined with the use of a standard leading operation. Go to step 1. Step 5: Define the solution of problem (11.1) - (11.4): AT
N fc=l fc=l
where z*k is an optimal solution to the Z-problem. Terminate. §11.3. The decomposition method scheme considered above has been ap plied for the multi-index transportation problem. 10 For example, the threeindex transportation problem takes the form: nTl3 3
Tl2 U2
inl il
m m 5X3l 5S 3 5S 3 ^11213^112^ c llt2t3 :r nt2t 3 ~> -* min, '
(11.13)
13 = l1 2122 == 11 1 1ll=l =1 I3= ri2 Til
nn i
l
X X *hhi ^11133 = = aa ni 3 .53 53
713 713
22 = 1 2 1 = 1 22 = 1 i\=l ni ni
l»s = = Tl , n * i
3
^3.
v x x^2*z = a i > *2 = v ^ ,
13 = 1 i l = l 13 = 1 i l = l 113 "2
n3
nz
,Bi X) *MI*» = ***!' ' *»i l== 1M 53 5533 *Mi*a = °ii l '.
13 = l1 12 12 = = 11 i3 =
^ 2 :xn ii 22 ii 33 > >0 0 ,, ii=T~nl; ii = l , n i ;
»2 *3 == lM , n22 ;; i»s3 == M l , »^3.-
(11.14) L (1 14) (11.15)
(11-i5)
(11.16)
(11.16) (11.17)
80
Systems Optimization
Methodology
Suppose that n3
n3 Tl3
=
ri2 Tl2
ni n\
2_^i 2-*t ^ ^ ' C ' l ' 2 ! 3 ~~ ^t3/ X ll»2>3 13 = 1 12 = 1 1 1 = 1 13 = 1 t 2 = l i l = l
under conditions (11.15) - (11.17). Solving the XA-problem reduces in turn, to finding an optimal solution to the classical transportation problem of ni xn2 dimensionality. Thus, the fc-index transportation problem reduces to the linear programming problem of a general form and the sequence of (it - l)-index transportation problems. §12. DECOMPOSITION FOR PROBLEMS W I T H A BLOCK-STAIRCASE S T R U C T U R E §12.1. Consider the algorithm for solving LP problem with a block-stair case structure of the matrix of conditions 20 using the Dantzig-Wulf decom position scheme. Consider the problem max Cxn , AiXi A\Xi
AJXJKBJXJ-!, AJXJKBJXJ-I,
j = =
Zj>0,
22 ,, . . . , n ,
(12.1)
j = l,2,... ,n,
where xx are variable vectors, dimensions of matrices A„ Bn (j = T~H; ji = 2~R) correlate well. Construct the problem dual to (12.1) minfe-U! , UjAj
}+lBj > uJ+1 Bj
Uj > 0 ,
,
j = 1,l n, n- - 1l ,
(12.2)
j = 1,71.
Assume that problem (12.1) has a feasible solution. We set forth the al gorithm for constructing an optimal solution to problem (12.1) by the given feasible solution. The algorithm comprises n steps. Let x°,... , x°n be a fea sible solution of (12.1). The algorithm step it results in constructing feasi ble solution (1) xl... , i j , that x) = xk, - = M = f c ; * } , j = n - J f e + 1 n
Dantzig- Will} Decomposition
81
constitute the optimal solution of the problem An-k+\Xn-k
max Cx Cxnn , _k+1+x°\X ,... +\ < < Bnn-k n_nk_ k, . . . ,,
•* *• • Ji St-Tl^n J*nXn
_ _
(12.3)
-^71 ^-*n — —1^-n—1 1^-71—1
as well as the optimal solution tij, 7 = n-k (12.3).
+ l,n of the problem dual to
§12.2. Describe in detail separate steps of the algorithm and their connec tions. Each step involves a finite number of iterations. Step 1:
Solve the problem max Cxn, ■ft-nXn 2l £>n%n~i
i
X„>0. £„ > 0. Let xn be a basis solution of the problem, and un a solution to the dual problem. Let x1 = x°x° j l , n n - 1 , \ = xn, u" = un and go to step 2. Before formally describing step 2, we formulate the problem to be solved at this step: max Cxn, yln-i^n-i
_; ^ n ^ n — 1 j
(12 4) (12.4)
i ^> n ^T n_ i .r„ ^ n —1 *J ■
We apply the Dantzig-Wulf decomposition method to solve this problem as well as a similar problem at step k. Let arLi.^n ea f feasible solution of (12.4). Assume that all the polyhedral sets under consideration are bounded. Let z ^ l j (i = T;N) be all the extreme points of the polyhedron An-xx^x < £ n - i < _ 2 - Its arbitrary point can be represented then as N
E
i=l
JV
i =i
Z
y
n-lXn-l'
i,»
82
Systems
Optimization
Methodology
(Here the first upper index belongs to the step number.) We shall present the point xn_x in a somewhat different way N N
E
1=0
N
where i n '_i = 4 - n E
2
i=0
l,i 1,1 l,i X n-l n-lI
Z
n-i = 1
or N
N N z2
X
/_*, ZZn™-1 '~1 *Xnn-i- l = X ^ nn_1- l -~ 2_^ 2 - / n'-l\ n ' - ln-l ( « - l ~ n-l) n - l ) '> 2_^ i=l i=l X
_
x Z
= =
'
Thus, the coordinating problem takes the form max CCx x nn, N N AnXn
— X + -D Bn n 2_j n-l) + 2_^ Z^n-\\ n ' x--n.-\ ll^n-l ~ n-\) i=l
— Xn-l n-1 > (12.5)
N N
1=1
Z > 00,, Xnn >
4 - 1 > 00,,
JV «t ==l ,U V
The vector xn = x\ is the basis solution (5), u\ - its corre sponding vector of dual variables. The optimality criterion for a feasible solution is constituted by fulfillment of inequalities.
A< = (*i-i - a=ili)«i > 0, i = \,N or min_ A; = min (zj,_i - Z n - i ) u l > 0 . i=l,JV
i=l, N
Testing the latter condition by the Dantzig-Wulf method can be reduced to solving an LP problem of the form max xn-iu -\uln, A < n-ixn_i n^x\_ •An—l^n—1 S B ^n— lXn2 X Xnn-1_ i > 0 .
— 2, '
(12.6) (12.6)
Dantzig- Wulf Decomposition
83 83
The objective function value obtained is compared to xl_tu\. Continue describing algorithm steps. The number enclosed in parentheses corresponds to the number of iteration at the step. Step 2(0): Obtain the optimal solution * J £ , of problem (12.6) and the dual variables u 1 ' 1 ,. If xh\ul < x1 , « ' , then let x2 = x1: i = Z «i_x = ^ " a n d go to step 3, otherwise return to step 1, i.e. continue solving the coordinating problem (12.5) and introduce into its basis a new vector (a£_j -x1^). Return ing to step 1 results in solving problem (12.5) x\, z*-i and ob taining its corresponding dual variables < - \ v1^1 (v1^1 is a new scalar dual variable associated with the constraint £ z£-i = 1). After the jth. return to step 1 go to iteration 2{j). Step 2(j): Obtain the optimal solution a ^ j 1 of the problem max x-n-iu]^3 , An-xX^! < 5 n _ i l J l _ 2 , Xn — 1
and the dual variables corresponding to x ^ + 1 . If X
n-1
U W
cu — — x■n-l nn- l u n
nn
T ^„
,
thenseti?=a;J,i = l , n - 2 ; 3 x
=
n-l
X
n-1
—
2_r ^n'-l\Xn-l
~ ^n'-l)<
j=l
x
« —xn'
w
n
—
a
> (tn-l
n
"n-1
'
Go to step 3. If X
n-\
U
n
> xn-\un
+ vn
>
then return to step 1 and insert into the basis of problem (12.5) the column ( x ^ - a ^ j 1 ) . As a result of performing the
84
Systems Optimization
Methodology
iteration in (12.5) we have xn, Z^-n i = 1,... ,j + 1 and the dual variables < J + 1 , v^+1. Now turn to iteration 2(j + 1). Evidently, the number of iterations at step 2 is finite and limited by the number of basic solutions (12.5). Step k(0): Obtain the optimal solution xkn~_Y+i of the problem max X n-fc+iti^^a, Xn-k+iu^zl+2> An+l-kXn-k
+l < B 5 n-k+lX _ f c + nl _Xk n _ f c , Xn-k+l In-fc+1 > 0
and the dual variables z * ! ^ corresponding to 4 - l + r
If
_k-l.l X
,,*-! < ~k-l fc-1 L n - f c + l u n - f c + 2 — x■ra-fc n-k ++l\uan-k+2 n - A : + 2 ''
x) = = xkf\ then set i* "n-Jb + l
—
k 1 M^;; u) uj = = uti*" = n-k n-fc + + 2,n; Jj = T ~\ , Jj =
"n-fc + l"
Gotostep(fc + l). If )t-i fc-i,i k -fe—i i . _fc-i fc-i U 'n-it+l 7i-A:+2 *** I n - A : +Xn-k+l l " ' n -Ufn-k+2' c+2 ' n-k+lUn-k+2
a X
then go to step {k - 1) with regard to the inserted column
(**zi+1 - ^t-l+i)- After the j'th return to step (fc - ! ) S°
to step k(j). Prior to it, we have an optimal solution to the problem of step (fc - 1) * , , j = n - f c + 2,n; %Z\iv i = 1,3 and the dual variables u3, j = n - k + 3, n, u * I ^ 2 , w * I ^ a . gtep fcjfl: Obtain the optimal solution x^^1 of the problem max x n -k+l u B -*+2 i, < -SBn„_t ^j4n-ifc+i£n-fc+i n - i t + l^n-fc + l < - k++ia; l £ nn_~kJ t ,, Zn-k+l > > 0 Xn-k+l
and the dual variables u^Uf
corresponding to x ^
1
fc-l,j + lk—l,j k-l,j+l k—l,j , , fc-l fc-1 fe— k—1,3l,j ,fc—1, , k-l,j j U U Un-k+2 — XXn-k "+" VVn-k+2 n-k+l n-k+2 — n-k + + llUn-k+2 n-k+2 "+" n-k+2 ' '
X Xn-k+l
then set xk = Xj Xj, , jj== n-k n - k + 2,n; 2, n; xkk == x)~ xk~ll, ,jj =
l,n-k;
j k r x x n-fc+l
- Tk~l -'S^r*-1'* — x n-*: + l 2—t ZZn-k+l n - k + l — ^n-k + l 2—t n-k+l
( k-1 \xxTn-k+l \ n-k+l
_ TX f c _ 1 ' ' "\ < Xn-k+lJ n-k + lJ <
If
Dantzig- Wulf Decomposition
85
u^=Uj, i = n — k + 3,n; aw
n-k+2 n-k+2
a — "n-fc+2' n-k+2'
aa
n-k+l
~- uunn~k+l ~k+l
'
go to step (k + 1). Otherwise go to step (fc - 1) and add the column (tf£lfc+, - a ; ^ ^ 1 ) to the matrix An — k+2-
Evidently, the number of iterations at the step k is finite and restricted by the number of basis solutions to the problem at step k. §13. SOLUTION OF T H E INTERVAL P R O G R A M M I N G PROBLEM §13.1. Consider the interval linear programming problem max Cx + b~
(13.1)
where A is an (m x n) matrix, h~, b+ are m-dimensional vectors, c is an n-dimensional vector. The interval linear programming problem can be interpreted as the Cp general problem provided the two-side constraints are regarded as a pair of individual conditions, and each free variable is represented as a difference of nonnegative variables. This procedure, however, enlarges the problem dimensionality. On the other hand, the interval programming problem finds numerous applications, say, in production planning, i.e. special methods for solving the problem are of interest. Describe an approach to solving problem (13.1) based on results 2 using the Dantzig-Wulf decomposition method as the base. Introduce the following notation: Smxn - an m x n matrices space, mxn with the rank r, R(A) = {y 6 Rm\ y = Srnxn _ & s e t o f m a t r i C es in S n Ax for a particular x € R } - the range of A matrix values, N(A) = {x G Rn\Ax = 0} - the A matrix zero subspace, M = {x € Rn\b~ < Ax < b+}-a set of feasible points for problem (1). Definition 1: Problem (13.1) is called feasible if M # 0. If A 6 S £ x n , then problem (13.1) is feasible with any b~ b"< since here R(A) = Rm.
86
Systems Optimization
Methodology
Definition 2: Problem (13.1) is called constrained if M ^ maxCr
0 and
Lemma 13.1: Let M / 0 . In this case problem (13.1) is constrained if and only if (13.2) Cx = 0 WxGN(A). WxGN(A). Condition (13..2)an nb tested if there is tht basis N{A). The following lemma provides a procedure for constructing the basis. Lemma 13.2: Suppose A £ Smxn
and E € S™Xn satisfy the condition
\ ^ ( m — r)n /
where p is a permutation matrix. Then the rank of (A) = r and that of matrix column
*-?{£) form t h t basis N(A). In the following lemma it is established that problem (13.1), for which (13.2) holds, can be transformed into the interval problem with a full rank, i.e. with the matrix A 6 S™xr. L e m m a 13.3: Let A <= S™xn and D e STrXn be such matrices that R{DT) = R{AT). Then ADT e 5 r m x r - Here the proof is evident. Lemma 13.4: Assume that problem (13.1) has the matrix A € S™Xn and (13.2) is fulfilled. Let D be the matrix satisfying the conditions for lemma 13.3. Then an optimal solution to problem (13.1) is represented as x* =DTy* N(A), = DTy* + N(A),
(13.3)
where y* is any optimal solution to the interval programming problem with the full rank matrix max CDTy, maxCDTy t T
(13.4) +
b~ < AD y < b
(13.5)
Dantzig- Wulf Decomposition
Proof: Using (13.2) and the fact that R(AT) optimal solution to problem (13.1) is equal to
87
= N(A) we find that an
(13.6)
x* +N(A), x*+N(A), where x' is any optimal solution to the problem max
Cx,
b~
b+
(13.7)
xeR(All).). xeR(A From the condition R(DT) = R{AT) for lemma 13.3 it follows that any element x 6 R(AT) can be presented as x = DDlyty
(13.8)
for a particular y £ RT. By substttuting (13.8) )n ((1.6) )nd droblem (13.1) in (13.4) we obtain the desired result. The following lemma will be utilized in constructing a decomposition procedure for problem (13.1). L e m m a 13.5: Assume that the interval programming problem is pre scribed by the matrix A £ S%xn. Its optimal solution is l x*=A-1w\ x*=A~ w*,
(13.9)
where
( >0
(h Wi Wl
= = II 6^ eiai + + (1 - Bi)bi di)bi ,
if {CA{CA-% )1= = \I = = 00 l
(13.10)
[<0
[a, z» = « Tl,n^ a and n d QO<<00i,<
Proof: Substituting z = Ax in (1), we obtain the equivalent problem maxCA-'z, maxCA-'z. b~
88
Systems Optimization
Methodology
the optimal solutions of which are prescribed by (13.10). A backward sub stitution gives (13.9). §13.2. At this point we will show that any constrained problem of interval programming is transformed to the equivalent decomposable form max Cx, b-
(13.11)
h-
where, by lemma 13.4, we can state, without violating generality, that Partition the problem constraints as follows 6f < A\x Aix < b'l, b~[ , < A2x < b 6+ b2 < 2 ,
(13.12)
bj < ^3" < A3 x < < b% H -, SrTXT,
Xr
where Ax G A2 e S« r A3 € S<m~T~^xr Here xj. saninnvertible matrix, and the submatrix A2 has a full rank. To be noted is that q is denned in an ambiguous way, but as shown below, q should be taken as large as possible (q = 0 can always be taken). It is possible to introduce new constraints of the form 64 < Bx < b+ , without violating the conditiont of (13.12) and ( ^ 2 n s STrx( The original problem (13.1) takes the form max Cx b^ bi < A\x < bf , b2 < A2x < b+ b2 ,, bbj3 < < A3x < bf , &7 < Bx < 66 j| .
(13.13) (13.13)
Dantzig- Wulf Decomposition
89
Note that x* is the optimal solution of (13.1) if and only if (x"tym) is an optimal solution to the following interval problem written in the decomposable form:
ft>(* «U(;)*(*,)■ \
3 /
\
3
m-T-q/ m-T-q I
\b3
)
§13.3. We set forth the decomposition method for (13.1). The decomposed interval problem (13.11) is written as the equivalent formula max Cx, (13.14) 6- < Ac A t < 6+ b+. . 6" To solve (13.14) the Dantzig-Wulf decomposition principle is applied, the coordinating problem_being constructed on_the equality x = x. Assume that 5 = {xe R?\b < Ax
N f). ^ f i = l, 2 t > 0 , i = l,N i=l
The same is true for the polyhedron
5S = = {xeR { x epf\bip|6-
Write the coordinating problem for (13.14) N
max 2_\{Cxi)zi, max>J(Cii)zi, ii=\ =l N 29
2V N
V f i f i - V & £ = o, tfei =l
«=1
i=i i=l
i i=l =l
Mi >>00, , »i == 1,7V, & lJV, llff>>00,, i= i= 1,7V. 1,7V.
(13.15)
90
Systems Optimization
Methodology
Problem (13.15) is the LP problem with many variables and unknown columns. According to the Dantzig-Wulf scheme, iteration of columns is performed by solving local problems. Introduce into consideration the (p + 2)-dimensional vector of Lagrange multipliers ((X A1i,...,A , . . . ,pA ,er,a r ), ( T 2 ) = = ( A(A,a , ( T 1 1, ,a )• p , e2 ( T2 2 )•
The vector (x\ 1,0)T or (-x\0,1)T can be introduced into the basis if it has a nonnegative estimate A; with regard to this basis A ti = (A,
l 1 \ -Cx
1 = (X-C)x +a1
il f\ = (A,
<0
=-A£'+or2<0.
2
According to the simplex method, the vector with a minimized estimate A is introduced into the basis A = min{ min min_((A ((A - C)xl + + c^), min (-Ax 1 + a2)} i=l,N
i = l, N
If A > 0, then we have an optimal solution. Define extremal points **, x* such that (A - C)x* C)x* = = min_(A - C)x*, C)xl, i = l,N
= min (—Xx1). —Xx* = i=l,N i=l,N
Since min (A - C)xl = min (A - C)x , i=i,fi
*i eeM«
then, using lemma 13.5, obtain x* = 2Y^J bfa ~bt^t + y* Yl b^ l a,,iek+
w h e r e i - 1 =(au,..
i£k-
,ap),
k+ = {i|(A - c)at > 0} , jfc+=
Jr - c)a 0} fc" = {i\{\ {z|(A-c)a <0} t t <
(13.16)
Dantzig-Wulf
Decomposition
91 91
Similarly we have (13.17) i£k+
iek~
w h e r e i " 1 = {au , ■ . ,ap), k+ = {i\ k+ {i\ - --&i o t >> 0}0
fc~ jfc- = {t| - \as \as < 0} }}
If A < 0, then either (£*,1,0)T or (x*,0,1) T are introduced into the ba sis according to which of these vectors has a smaller characteristic estimate. If A > 0, then we have an optimal solution to problem (13.14) in the form JU
•—
/
'-.-[*(.-
\
— i •(
/
<,
where Jb, = {i\zt > 0}, k2 = {i\zt > 0}. The initial basis is chosen in the following way. Consider the problem equivalent to (13.14) N
P V
max2_](Cx*)z max 2_](Cx*)zi — 9&/~\vi, 2_\vi i l i=i
i=\
. _
.
.
, .
.
,
(13.18)
(:: :')(:) - R)' ( ! : " : ) ( : ) ■ ( : ) ■
where v is the p-dimensional vector of artificial variables, D, D are matri ces with columns x\ x\ respectively, and 9 is a sufficiently large positive number. First find optimal solutions to the subproblems max Cx, maxCx,,GS. m axCx,
x £x£S, S,
(13-19) &M)
,GS.
If x, x are solutions to problems (13.19), then the initial basis is taken in the form
O'fi)'"© *(!)•
( ! ) ■
( ! ) ■ ' ( : )
* ( ! ) ■
<1320>
'
(13.20)
92
Systems Optimization
Methodology
The plus or minus sign in the column depends on the component sign (Xi-Xi),(i = T^). §13.4. We set forth the algorithm of the above procedure for solving the interval problem LP. Assume that problem (13.1) is reduced to the decom posable form (13.11). Step 0: Finding the initial basis. To do this, subproblem (13.19) is solved and the initial basis B is taken in the form (13.20). Thereupon the basic feasible eolution (*, z,v) and iti corresponding dual variables (A,(Ti,(72) are sought. Step 1: Calculate the value (A - C) and with formulae (13.16) and (13.17) determine the extreme points x*, Z* delivering a minimum to the relative estimates A, A of the relative basis vectors (x,l,Q)T, T (x,0,1) . Let A = min(A,A). If A > 0, then the basic solu tion under study is optimal. Go to step 4. Step 2: One of the vectors (S,1,0)T, ( f , 0 , 1 ) T possessing a smaller rela tive estimate A, is introduced into the basis of the coordinating problem. Step 3: The vector introduced into the basis is multiplied by B " 1 and, with the use of the standard leading operation, new inverse matrix and vector of dual estimates are defined. Go to step 1. Step 4: Define a solution to problem (13.11) l x* — 2_, zZi^ i%x 1++z2z2 ^***"' '
where zu i, is the solution of the coordinating probleme Terminate. §14. EXTENSION OF THE DANTZIG-WULF DECOMPOSITION PRINCIPLE TO NONLINEAR P R O G R A M M I N G PROBLEMS §14.1. Since linear programming is an adequately developed means for solving problems, it is natural to try to use its methods in nonlinear prob lems with the help of linearization.
Dantzig- Wulf Decomposition
93 93
Consider the mathematical programming problem min
M
gi{x) < 0, 9i(x) < 0, where / , g, zre eonvex functions, x G Select a set of mesh points xux2,... has an element of the form T
f(x), fix), (14.1) ' (14.1) i = l,m, i = l,m, EnG ,xT. The convex hull of this set
T
x = ^2ztxt,
}_izt = }_zt = l, 1, zt>0,
t1=1 =l
t=l,T
1=1
Replacement of / , 5 t (i = T^n) by their linearization on the lattice leads to an LP problem with respect to the variables zt: T
mm^2ztf(xt), t=i T
^2 ztgi(x zt9i(xt) < 0 0,,
i = Tjn, l,m,
(14.2)
t1=1 =i T
Y^zt
= l,
zt>0.
1=1
Obtaining exact approximation of the nonlinear functions / , 9i (i = T~^) calls for many nodal points. Therefore it is desirable to refine approxi mation only in the neighborhood of a minimum, and it can be done by employing the simplex method for choosing new mesh points when solving appropriate auxiliary problems. Assume that at the beginning of the iter ation we introduce ehe eattice with hhe eodal points xu .,. .xTT Let t**} be an optimal solution to (14.2), and A^ (» = I~m), AQ dual variables. The current approximation of an optimal solution to problem (14.1) is: T X
= }
jZtXt.
t1=1 =l
From all the new nodal points that might be introduced for improving the approximation the simplex method would have chosen the point corre sponding to the column with the smallest characteristic difference, i.e. the
94
Systems Optimization
Methodology
point xt which minimizes m
m i—l
Calculating that point reduces to solving the nonlinear problem without constraints k (14.3) mmlf(x)-^2\ min {/(*)-]£ (14.3) gi{x)\*?#(*) I The problem can be solved by the proper unconstrained minimization method. If the minimum iisttained dathe eoint xk, then nroblem (14.24 is supplemented with a new column (g (gil(xk),...,gm(xk),l)
pk =
and f(xk)zT+i is introduced into the objective function. Thereupon prob lem (14.2) is newly solved. If the objective function and constraints in (14.1) are linear and denned on the convex polyhedron, then any point in side it can be represented as a convex combination of extreme points form ing a finite lattice, and approximation can be made exact. The process of linearization on the lattice is then transformed into the Dantzig-Wulf decomposition algorithm for the LP problem. In this case retaining all the generation columns in the coordinating problem is not necessary. In the nonlinear case the existing proof of convergence- requires retaining all these columns. If £ ] \lgl(x) {x) > \ - Ag \k > 0, A, < 0, i = 0^i min I| f(x) - Y Q,ra
(14.4)
then the current approximation is optimal for problem (14.1), which is shown in the following theorem. T h e o r e m 14.1: Let (A*,Ag) be the dual estimates corresponding to the optimal solution of the coordinating problem (14.2) on iteration k. If (14.4) is fulfiiled, ,hen nhe point txk £ ] k t x t is optimal, i.e. t
min
ff;(x)<0, i = l , m g;(x)<0,
f(x) =
f{xk).
Dantzig- Wulf Decomposition
95
Proof: Characteristic differences for basic variables are equal to zero: 771
in
k
ft = f(xt) + Y,ukgt(xt)
+ uho = 0,
i=\
where u\ = -Af k > multiplying these equalities by z* and summing them, we get m 7TI
fc 5>*/(zt) + 5>* 5>^(* = o.o. 5>^f(x*)+5>*5> &(x()+wS t)+i4 = t
i=l
t
Under conditions for complementary nonhardness we have m m
5>*5>*§i(xt) = 0. i=l i=l
tt
= -zk-
Thus, «S = - E 4f(xt)
Consider the inequality min ££(x, ( x , u) u ) <<mmin i n / (fix), x), Xxz. tli xeE"
V
x t JKo ' ~ x€S
(14.5)
'
where S5 — = {x\gi(x) {x|<7i(x) < 0, i = l , m } , £(x,u) is the Lagrangian, £(x,u) = f(x) + ^ u ^ x ) . Since / is convex i
and xk are feasible, then
k k k m iinn/ (/ x( )x )<
(14.6)
Add the value uk + zzkqual lo zero oo the rightthand sides sof14.5) and, considering (14.6), we obtain kk minn {C(x,u mm{C(x,u )) n x&E x£E
k fc + uk]0} + zk <mmf(x < min/(x ) ) < zk . xes x€S
Under conditions for the theorem min {C(x,uk)
n x€E x€E"
+ uk)} > 0,
(14.7)
96
Systems Optimization
Methodology
so that this value can be removed from the left-hand side of (14.7) zk < m i n / ( x )
XtzS
i.e. mmf(x) xES
= f(xk)
= zk.
The proof is completed. §14.2. Consider the case when / , $ are additively separable, i.e. x 9i = Ylsij( 3Y i3 == 1M > n; ii == i,TO. ^2 9ij(XjY, M".
f = ^2f](xj); ^2 fjM' 3
In this case the advantages manifest themselves. A set of the nodal points {xjt} can be denned for each variable x} and each of the functions fh gtJ is linearized separately so that the problem arises min J^Zjtfjixjt), J2z]tfj(xjt), j,ti,t
y^ZjtgijjXjt) 7 1,, 2 j % S i j ( ^ t ) < 0, i1 ==1 l, 7, m
i,t Y^zjt = 1. 1, 3 = hn; !> n ; t
(14.8) (14.8)
Zjt>0. zjt>0.
Auxiliary problems (14.3) are divided here into independent singledimensional problems min I fjixj) fjfa)
5 ^ h9ij($j) k9ij($j) >\ - Y^
§14.3. We will set forth a nonlinear variation of the Dantzig-Wulf decom position principle.34 Consider the problem min = ^^2f/ li(x( lz)\,i ) \ , m n Jlf(x) /(z) =
(14.9)
Dantzig- Wulf Decomposition
97
where ft (i = T~p) are convex functions, and linear connecting constraints are specified vv ^ArXi = b, (14.10) J2^x = b, (14.10) i=\
t
and a series of possibly nonlinear constraints is given with respect to each Xi: (14.11) xtteS€l, Si, i = Tj>, l,p, where 5, are convex sets, not necessarily polyhedrons, A% (i ~ T7p) are m-dimensional matrices. Inasmuch as the constraints on each x> are inde pendent, it is possible to choose an individual lattice of nodal points for each of them. Each function ft(xi) is substituted by its linearization on the lattice {£<}: (14.12) fl(xt) = ^2ztJl(xtl), (14.12) * where J ] 4 = l, A > 0 .
(14.13)
t
Its associated values of
follows xi = YJ4*i-
(14.14) (14.14)
t
Write the approximating linear problem for (14.9) - (14.11): r m m i n in jz Z= 2= ^£* 4/ i/( ^* *) )L? i
(14.15)
^2(Ai4)zt J2(A iXM = b,
(14.16)
5534 3 4 = 1, i=T^p~, i-l,p,
(14.17) (14.17)
*f>0. z\>Q.
(14.18)
i,t
t
Inasmuch as x\ € Sz and St is convex, the point x{ in (14.14) is also an element of S*. If z\ satisfies (14.16), then xt sattsfies ((4.10) and represents a feasible solution to the original problem. As before, it is undesirable to
98
Systems
Optimization
Methodology
select many nodal points in advance. It can be avoided with the use of the simplex method for generating them as needed. The column related to z\ is T T (14.19) p\ = {A{A,xlh) lxlh) Assume that a feasible basis in (14.15) - (14.18) exists, and let (Ai, A2) be dual variables. Then the characteristic difference corresponding to the column p\ is this: Ai = fi(xi)-\iAi2*-k At=M4)-hAn4-\ (14.20) 2i.2i. The problem of determining the point of x\ £ 5 „ transferring a minimum to this expression, takes the form min{/ xah^/ffo) M u x i ) ,, i (x 1 ) - XiA.x,}
(14.21) (14.21)
x, £ Xi € S„ 5,,
(14.22)
ii ==I ~I,p. p.
Let x, be a solution to the subproblem 1. Assume that if only one of the characteristic differences is negative: fi(xi) fi{x - XiAiii X\Alxl ~—\2l\2,< <0 0 for for anani. i. t) —
(14.23)
Form the columns pPi {Atxl,ll)T, = (Atii,lif, x =
(14.24)
ii = — hpI,p.
The columns are utilized in constructing a reduced coordinating prob lem. The problem has, as variables, a set of basic variables {z*} plus p new variables zt, one for each column p,:
min zz
\ = £ £ *«/«(**)+51 4/i(«!) + Y,*ji(*<) *»/*(*<)[.\ ' i
I
I
tba,
)
]jT ^ ( ^ x f H + J£(ji«*i)fj = b, Ji
Uo.
i
l ^ f + I ^ t = l, i = hP, »J>0, £»>0. tbaz
i
The problem has as the initial feasible basis associated with the variables z\ and p columns p\ of form (14.24) as nonbasic columns. §14.4. We set forth the algorithm of the nonlinear analog of the DantzigWulf decomposition method under assumption that the coordinating
Dantzig-Wulf
Decomposition
99
problem (14.15) - (14.18) has (m + 1) constraints in the form of an equality, i.e. conditions (14.11) are regarded as a single whole. Step 1: Obtain linearly independent solutions to i | € S„ t = l,m + 1 and calculate the following values v
i = T~p,
p
at = 'Y^Alx\,
7t = ]T]/i(Si)i t = pi,... Pi,--.
i=l
,Pm+i,
i=l
where (Pl,... ,pm+i) problem.
is the current basis of the coordinating
Step 2: Obtain the dual variables (Ai,A 2 ) from the formula r\ r\
\ \ I
api a p
Q Q p p 2 2
i
Q Q
'"
P'"+i P™+I \
1 1 1
1 /
=
/i
\\
^7w'"' '7p'"+^ '
5*ep 3: Obtain solutions to the following p problems min A; i A ^ ; x t x€leS A, = f/l,{x( i!), ) - A AiAiSj; 5 tl
(14.25)
Denote the optimal solutions of these problems by xx (i = T7p) and the corresponding A ; by A, = fi(xi) fi(xi) - Ai AiAiXi, AiXi, ti = l,l,p. p.
Step ^: Calculate pp
A* = ^
A, - A22 .
i= = l\
If A" = 0, go to step 6. Step 5: Calculate P v
a = ^2Alxi; (X
:=
7
i=l
A{X{
vP
\
^2,fi{xi), 7l = £/*(&), t=l
introduce the column (*) into the coordinating problem and carry out one operation of the simplex method.
100
Systems Optimization
Methodology
Let Zi,i = l,m + 1 be the current solution of the coordinating problem. Go to step 2. Step 6: Calculate the optimal solution of the original problem (9) - (11): X — — \Xi,... \X^, . . . , Xp) Xp) , Pm + l1
where x* — JZ zztx\, ti\, t=pi
i i—T^p. — \,p.
Terminate. In conclusion it should be noted that convergence of this algorithm to an optimal solution of problem (14.9) - (14.11) was proved by Sekain.34
Chapter 4 PARAMETRIC
DECOMPOSITION
§15. KORNAI-LIPTACK M E T H O D §15.1. Consider the linear programming problem max ex, A°x
(15.1)
x>0, where c, x are n-dimensional vectors, b° is an m-dimensional vector, and A0 an (TO x n) matrix. Partition the matrix A0 into submatrices
A° =
(A° (Al...,Al), 1,...,Al),
where with each j £ [1 : k] the matrix A? hasTOx n^ dimensionality. Then partition also the vectors c, x respectively: c ==(( cc ii , . . . , C C c ff cc ) ,
X =
{Xl,...,Xk),
the dimensions c^xjij = M ) being equal to n,. With due regard to partitioning problem (15.1) is transformed to the form kfc
max > J CjXj , j=i
101
102
Systems
Optimization
Methodology
(15-2)
U
(15.2)
i=i Xj > 0,
j =
l,k.
Introduce the m-dimensional column vectors y3 = (j - MO satisfying the condition (15.3) Formulate k problems LP: m a x CjXj ,
A)x3 < AjXj S y Vj3, >
(15.4)
Zj > Zj > 0. 0. Introduce the vector y = (pi,?.? ,yk) and denote by My a set of the vectors j/ such that (15.3) is fulfilled and problems (15.4) have solutions. Optimal values of the functionals of problems (15.4) are functions from the values of yf
fi = fiiVi)fiiVi)Then problem (15.2) reduces to the following problem k
m a x ^ y ) = 53/j(|»j), maxF(y) ^fjiyj), k
1=1
(15.5)
3= 1
If decomposing the matrix A0 is interpreted as its partitioning into k subsystems, and the vector 6° is regarded as a common resource of the system, then problem (15.5) resides in finding optimal distribution of the common resource. Here the constraint on the sign of the values ys, j - 1, f is not presupposed since the subsystem can also consume resources. The analysis of problem (15.5) is complex in that the function is specified algo rithmically. To find the values of this function we have to solve LP problem (15.4). Application of one or another scheme, when maximizing the func tion ?{y), generates an appropriate decomposition method based on the Kornai-Liptack decomposition principle. In the initial work by KornaiLiptack,14 problem (15.1) was reduced to the maximum problem and to
103
Parametric Decomposition
obtain an optimal solution it was suggested to employ the Brown-Robinson iterative scheme. But, as noted in Ref. 28, numerical experiments show that the convergence rate of the Brown-Robinson method is basically un satisfactory, which results in its modification and construction of essentially different schemes. §15.2. We set forth the iterative scheme for solving problem (15.5) based on application of the method of feasible directions in the set of the variables. 28 Return to problem (15.5) where the original problem (15.2) takes the form k
max2_] c j x j >i (15.6) (15.6)
0 Y,A»x y2A°xj , 3
3=1 / l 7 JU *j ^ ^
^7
}
XJ > 0,
j~l,k,
where with each j = 171, b3 is an m r dimensional vector, A, is an (ms x n3) matrix. As before, introduce the m-dimensional column vectors y3jj = 1, c, and formulate it problems LR m a x CjXj , A3jXj Xj < Vj y3 ,
~
'
(15.7)
/ I j3X j3 ^— Oj3 ,'
iXjj >> 00 .
In Ref. 29 it has been shown that the functions f3{y}), j = Ijfc are concave and piecewise linear, and break points correspond to replacement of the basis in problems (15.7). Consider the scheme of feasible directions for the case where the unique optimal dual variables Uj, j = Tjs correspond to some given vectors yjo, j = Ijfc. By the first duality theorem, we have m
m
j
+I /i(%) = iE 4 4 + ^ ^ = E W ' =l i=l i=l
i=l
15 8 (15.8)
( -)
104
Systems Optimization
Methodology
According to the scheme of the method of feasible directions, subsequent approximation is defined with the formula az »Vh i i ==t fVJO i o ++a * ij> »
k J3=- l.1.fc>i
(15.9) ( 15 - 9 )
where z, is the m vector of feasible directions_satisfying a particular con dition for valuation, e.g, - 1 < z)< zj j = l,fcl c = T~fH and a is the nonnegative parameter, the magnitude of the step along a direction. Substituting (15.9) in (15.5) with regard to (15.8), we obtain the fol lowing LP problem: k k
m m
}=11=1 ] = 1 1=1
k
*
Y^ 2 ^ zz)j <<00,,
it = = T/m, l,m,
(15.10) (15.10)
i=i
- 1i < ^ < 1, i,
i = I7fc, l,fc,
ti == T7m, l,m,
which falls into m independent subproblems of the form k
maxY]
u)z},
J3=1 =I
- i < 2* < i,
= TTfc i,k ij =
the solution of which is evident so that ' -"1I, . jj = = ' 2h,, 2* z) = < i, 1, ij = = hft,, .
0, j^h,l22,
i = l,m,
where the indices J lf Jj are prescribed by the conditions ujj = max M* ,
«}2 = min MJ u lj .
Thus, optimal rearrangement coincides with that flowing from elementary economic considerations; a resource is removed from the subsystem, for
Parametric
Decomposition
105
which it is least valuable and transferred to the subsystem where it is most valuable. To complete the description of the algorithm, indicate distinctive characteristics of selecting the a parameter determining the length of step along the direction selected. We turn to problem (15.5). The magnitude of a is defined as a = min {ax,a2}, where Q l is the maximal value of a satisfying the constraint of (15.5) kfc
kk
ji = i
>=i
where a2 is derived from the maximization condition for the single variable function A: k
^(") = £/;(»»), J=I
AjXj < yj0 + azj, AjXj < yj0 + azj,
AjXj
x3 > 0, x3> 0,
j = 1, k. j = 1, k.
Consecutive application of the scheme of the method of feasible direc tions, as shown in Ref. 31, ensures monotone convergence to an extremum for the finite number of steps. §16. S O L U T I O N T E C H N I Q U E FOR BLOCK-SEPARABLE N O N L I N E A R PROBLEMS §16.1. Consider the following block-separable problem k
min 55 JJ /»(*«) > t=i
* ^9i{?i)
(16.1)
k
< b,
jI— * = J.l X{ cc O ti j, X^
2 %— — 1XjK , ft .*
Rewrite (16.1) in the equivalent form, introducing the m-dimensional vectors Vi(i = hk)
106
Systems
Optimization
Methodology
kk
max^ max^
fi(xj), fj(xj),
i=l
gi(xi)
ii = =
l,k, l,k,
(16.2)
kk
i=i CC{
% —- J . j K .
i^i,
KZ
With the fixed » = { » , . . . , » ) problem (16.2) falls into ifc local problems of the form min/i(a; m i n / ; ( x jl)),, (16.3) V (16.3) It is natural that ft should be chosen so that subproblems (16.3) might have feasible solutions, i.e. ^yi£V V ,t
(16.4)
= {{yVlt\3xt S S; Si :: s,-(x < 3/,} 9i{xi)t) < Vl}
Introduce into consideration the functions Wi(yi) = mm{fl(xi)\gi(xl)
< yu
Xi € Si},
i = 1,k.
Evidently, to solve the original problem as a whole one should choose yt minimizing the sum of these functions. Thus, the original problem (16.1) is equivalent to the following coordinating problem k
w mmw{y) mmw(y)== J2^VJiifli), l^), it = = li
k k
b J2y*j ) » i < 6>,
(16.5) (16.5) (16.6) (16.6)
»=i
(16.7) i = l,k. m€Vi, i = ljfc. (16.7) For LP problems such an approach to solving the original problem is described in the preceding paragraph and is called the Kornai-Liptack method. Introduce the following propositions: V% € Vi,
(1) Si are nonempty, compact, convex; (2) fi(xi) are convex and differentiate on 5 t ; (3) Each component of the vectors gt is convex and differentiated on S;; (4) Problem (16.1) has a feasible solution.
Parametric
Decomposition
107
Evidently, the convexity of the set Vi and the function wt on the eet follows from the convexity of fl,gl,Si so that the coordinating problem (16.5) - (16.7) constitutes the convex programming problem. The coordinating problem feasibility follows from the feasibility of the original problem. From compactness of S; together with continuity of fi,gl, it follows that the ith subproblem has an optimal solution since it has the feasible one. liy° = (y°,...,y°k) constitutes a solution to the coordinating problem (16.5) - (16.7) and x°z is a solution to problem (16.3) with yt = y°, then x° = ( x ° , . . . , x°k) is a solution to the original problem (16.1). §16.2. As in the preceding paragraph, we will employ the Loitendake method of feasible directions to solve problem (16.5) - (16.7). We begin solving the coordinating problem with a feasible vector y, then construct the problem for determining the direction S or establishing the optimality of y. If y is not optimal then carry out a step in the direction 5, determining a new distribution, and the process is repeated. Denote
T-
< 2/lJ^t/t < b, yx £ K,
i = l, k \
Assume that a feasible point of y = (Vl..,., yk) € T is known. To perform a small step in the direction St from the point yt here yl + aSi € Vu suppose that y{ € int Vt. Let S = ( S i , . . . , S&) be the assumed direction of search. To be useful, S should possess the following features: (1) to be feasible at the point ofyGf, y + aS aS€T, € T,
i.e. e > 0 must exist such that 00<
(16.8)
(2) the feasible direction S has to be suitable for w at the point of y € T, i.i. w(y + aS) aS)<w(y)-ae, < w(y) - ae, 0 < a < e.. (16.9) Note that the parameter e for (1), (2) is the same. Thus, for S to be a feasible direction, it is necessary to fulfill the following conditions in a field a >0
108
Systems
Optimization
yi+aS Vi +aSi l€Vi,eVi,
Methodology
(16.10)
i = TJ l,fc
k
(16.11)
£>,+aSi)<& i=l
(16.11) is rewritten as k
k
(16.12) (1612)
aaj2Si
t=i t=l
Since under our assumption yi € int Vu then (16.10) is satisfied for any k
S direction. If the fcth component of the vector b - £ yi is positive, then i=i
there exists the region of such a > 0 that the fcth inequality is satisfied with any choice of the vectors {£} so that the corresponding constraints may not be considered. If the fcth component is equal to null, then a shift along some directions violates this constraint. Let fl = 0 | 6 j - ] T y t f = o0l .
s = jil&i-I>tf = }-
(16.13)
Then from (16.12) it follows that for feasibility of the direction one should fulfill the conditions k
s B J2 jeB 5 2 5a* J <SO, ° - j£ i=l
(16.14) (16.14)
Evidently, finding a locally better suitable direction requires finding the feasible direction that minimizes linear increment of an objective function, i.e. one has to solve the following problem k
mmAw(y,s) min Aw(y, s) = - ^Aw ^ A i{yw i,s1^),, s 4 ),
(16.15)
i=l k
52 J2 s5^« £SO, °> j3eeBB,,
(16.16)
i=l
N Il<<1i. . N
(16.17)
Parametric
109
Decomposition
Establish the optimality criteria for y and show that (16.15), (16.17) always yields a feasible direction, if any. Theorem 16.1: Let y be feasible for the coordinating problem. If s = 0 is an optimal solution to (16.15) - (16.17), then y represents a solution to the coordinating problem. Proof: Since Aw{y,0) = 0, then Aw(y,s)>0 > 0 Aw(y,s)
(16.18)
for all s satisfying (16.16). If w is additionally defined by the condition k
w{y) = oo, oo, «K») =
if if
^yt b' £b, Y,y^
(16.19) (16.19)
i= l
then for the . vectors not satisfying (16.16), there exists W(y + as) = oo, Va > 0 so that Aw(y,s) = oo. Thus (16.19) holds for all s G Em. There exists a theorem (say, Ref. 19) according to which y is the point of minimum for w if and only if Aw(y,s)>0,
Vs£Em. VseEm.
The proof is completed. Theorem 16.2: Let s be a solution to (16.15) - (16.17) for a particular y 6 T If Am(y,s) is negative, then s is a suitable direction for w at the point y. Proof: Since, according to the condition, s is a feasible direction, then a exists such that y + aseT, 0 0<
(16 20) (16.20)
There exists theorem (19) according to which if the function w{y) is finite the difference relation ,., ,XN w(y + Xs) - w(y) a(X) «(A) = = ~x
110
Systems Optimization Methodology
is nondecreasing for all A > 0. Together with (16.20) this statement leads to the existence of a particular e > 0, A > 0, thus a(X) < -e, 0
(16.21)
Denote 6 = min(e, A). Show that
a(\)<-6,
0<\<6.
(16.22)
(1) e > A. (16.21) implies Q(A) < -e < -A = -6,
0
(2) e < A. Due to (16.21) Q(A)
<-E
=
-6.
Since the latter holds for all 0 < A < A it also holds for all 0 < A < 6. Suppose A* = min(o:,<5). Then due to (16.19') (16.23) t/ + a s e f , 0 < a < A* and owing to (16.22)
a(X) < -A*, 0
w(y + Xs) < w{y) - XX*, (16.24) 0
f(z)>f{x)+x*{z-x), \lz£En. (16.25) (16.25) T h e o r e m 16.3: L e t £ ( x , u ) = f(x)+u(g(x)(x) — y) be the Lagrange function for problems (16.3) and x° the solution of these problems. Then — u° =
111
Parametric Decomposition
{—u0};; u° is the solution of the duality problem of (16.3) in the subgradient of w at the point y if and only if (x°, u°) is the saddle point of the Lagrange function, i.e. £(x0,u)<£(x0,w°)<£(z,M0). Proof: Necessity: Since (x°,u°) is the saddle point £(x,u),
then
f(x°) + u°(g(x°) -y)< (x) - y), V f{x) xeS + «°( 9
or, since /(x°) = w(y), u°igix°) - y) = 0, then (16.26) wiy)
If there exists x £ S such that gix) < z then for such
x, gix) in (16.26) can be replaced by z:
i»(y)
u°iz-y).
Calculate inf of the right-hand side of this inequality by x € S so that gix) < z. Then (16.27)
w(y)<wiz) + u°(z-y).
If there is no x € S such that gix) < z, then w(,z) = +oo. Comparing that relation with the definition of subgradient (16.25), we see that —w° is the subgradient of the function w at the point y. Sufficiency: Assume that (16.27) holds for all z e Em. Choose
hence
z = gix°), (16.28)
(16.28)
wiz) = fix°)=wiy).
(16.29)
Substituting (16.28), (16.29) in (16.27), obtain
u°igix°)-y)>0. And since u° > 0 and gix0) - y < 0, then u°igix°)-y) = 0.
(16.30)
112
Systems Optimization Methodology
Condition (16.30), combined with the relation g(x°) < y, yields two from three conditions for the saddle point. It remains to show that x° minimizes C(x,u°) ) on S. Substituting the equalities f(x°) = w(y), u°(g(x°) - y) = 0 in (16.27) obtain
( f(x°)+u0(g(x0)-y)<w(z)+u0(z-y), V* € Em . (16.31) For an arbitrary x € S set z = g(x). Since x is a feasible point for problems (16.3) with y — z — g(x), then
w(z)
(16.32)
Using condition (16.32) in (16.31), obtain f(x°) + u°(g(x°) -y)< f{x) + u°(g(x) - y), which proves the third condition. The proof of the theorem is completed. There exists a theorem according to which a convex function is differentiable at a point if and only if it has a unique subgradient at this point. Since, with x° being optimal for problems (16.3), (x°,u°) is the saddle point a if and only if u° is the solution of the problem dual to (16.3), then the following statement flows from theorem 16.3. Theorem 16.4: The function is differentiable at y if and only if the prob lem dual with respect to (16.3) has a unique optimal solution. Theorem 16.3 can be employed for obtaining the explicit expression Awi(yi, Si) at the points yx € int V;. Theorem 16.5: Assume that the convex function / has the nonempty region S of its finite values. Then A / ( x , s) = max{a;*5|a;* - subgradient / } , where x € int 5. According to theorem 16.5 there exists (16.33)
Parametric
Decomposition
113 113
where V% is t h e set of all optimal dual variables for problem (16.3). Let each set * be defined with the use of the vector function a si {xi\ci(x.Ui ++ I^.Xt Ig'Ui I^Xi=-Vfi, = -V/t,
(16.34) (16.34)
m(ft-tt)=0, m(ft-tt)=o,
(16.35)
xAlCli C t ==0, 0,
(16.36)
Ui>0;
Xi>0, At>0,
(16.37)
where Ig<, ICt are eacobian for ru a i.ei the matrices, s , e rows of which ara t h e gradients of t h e functions constituting gt,c,: Due to (16.33) Awiiyt,
Si)
= m a x { - « t 5 t | ( U t , A,) satisfies (16.34) - (16.37)} .
(16.38)
Since $ < yua < 0, conditions (16.35), (16.36) are equivalent to the following mj = 0, u^
if g^ gij — yij y^ << 00
(16.39)
X{j=0, Xij = 0,
if dj<0. Cij < 0 .
(16.40)
T h u s , (16.38) can be rewritten as
m ax < max J — - £V«J«UijSij * « [ ,>,
(16.41)
VC v 53 ^9ijUij A1:) == - V VgijUij++53 53 Vc^ <J'*« / /*i .,
(16.42) (16.42)
ij
3i « y >> 00,, Wjj
A Aij^ >O 0, ,
(16.43)
114
Systems Optimization Methodology
where summation is done by such j t h a t 9ij - Vlj = 0,
Cy = 00
Write t h e dual problem t o (16.41) - (16.43) with t h e provisionally fixed st: m i n { - V f/ ii ZZii}}, ,
(16.44)
VgijZi Vg > -Sij -s^ ,, zjZi
(16.45) (16.46) (16.46)
VCijZi>0. Vc^>0. Thus, for the fixed 5 = {su. ..,sk)
t h e directional derivative w(y) is equal
k Aw{y, s) = Aw{y,s) = Y^ ^ m imn i{n-{V~ /V! z/ ilZ| (i z| (1*, ,5 *I )) satisfy (16.45), (16.46)}
(16.47)
it = i
In t h e problem of direction selection we had t o find t h e vector t h a t was minimized (16.47), and this is equivalent t o solving t h e problem:
min <^ - V Vjxzx \ , | -f^Vfi*], it it kit
(16.48) (16.48) (16.49)
i1= 1l
1 i = 1l
VftjZ* > -aij, Vft,*,>-*i,
jj-.gM): gij(x°)
-yij= yij=0,0,
i=l,k i=TJ
(16.50) (16.50)
VcijZi VcijZi>0,> 0,
j ::ClJ (x°) ClJ(x°)
= = 0, 0,
= Tje. 1, k . i=
(16.51)
T h e problem constructed will be linear if we use the norm ||s|| = max|s_,|. \\s\\ i Then requirements for normalization are written as - 1 < Sij < 1 .
(16.52)
R e m a r k 1 6 . 1 : If all the functions w, axe differentiable, the problem of direction selection reduces t o t h e following (see (16.15) - (16.17), (16.33)):
Parametric
Decomposition
115
mini - ^ u ° S 5 i >, vI
22 S*>j 0, 3; €€ B B ,, J2 'J ^< °>
(16.53)
i=l
- 1 < ay < 1. Remark 16.2: Problem (16.48) - (16.52) has an angle structure, (16.49) connecting constraints, (16.50) - (16.51) angle blocks. It remains to con sider the problem of step length selection. If a suitable direction s has been found, then a new point y is found via selection of the step a > 0 in the expression yn+l = =yynn+as. We set forth one of the existing schemes for selecting a. Solve the convex single-dimensional problem min{w(y + as)\y + as € T} T}
(16.54)
To solve the problem, we can apply any single-dimensional minimization technique for the case of constraint absence when it is considered that the constraints y + as £ T in the explicit form appear as k
^(Vr Y2(y* + + asas,) i) <
(16.55;
i=l
, Vi, ytl+as +as1€V t l€
i=T7k. i = 1, k.
(16.56)
Constraint (16.55) is equivalent to calculating the upper bound of a. Vio lation of constraints (16.56) can be detected by the absence of feasible solutions in a subproblem. §16.3. We set forth the algorithm for solving problem (16.1) Step 1: Select y° so that 3x% : gt(xi) < j/°, t - U b . Set n = 0. Step 2: Solve the problem of the s direction selection for (16.48) - (16.51) or, in the case of differentiability of all Wi, i =1,k problem (16.53)) Step 3: Determine the value of the step a in the direction s from solving (16.54).
116
Systems
Optimization
Methodology
Step 4: Test optimally criteria. (Theorem 16.1) If s = 0(\\s|| < e), go to step 6. Step 5: Determine a new point y with the formula yn+l = =__nn+as. Setting n = n + 1, go to step 2. Step 6: Terminate. §17. ON P A R A M E T R I C DECOMPOSITION OF T H E RESOURCES ALLOCATION PROBLEM The preceding paragraph was concerned with one method of parametric decomposition for what might be termed the resources allocation problem. k
min^/i(xi), i=i kk
^giixt)
(17.1)
< b,
t=i X{
xz S{,
I — ~ 1, K ,
where xt are the ^-dimensional vector, gt the m-dimensional vectors {i = Ijfc), m < k. A choice of subproblems and a coordinating problem was suggested for the problem. Here we shall consider some other possibilities of selecting subproblems and appropriate conditions for coordination38 so that they might constitute a unified system equivalent to the original problem (17.1). The present paragraph is rather of a survey nature since the approaches to solve local and coordinating problems are not considered here. §17.1. We expand problem (17.1) into the following (t = l,...,Jb) mmfiixi), min/<(*)> gi{xi) + ai
subproblems
,am)i is the vector of coordinating parameters.
(17.2)
Parametric
Decomposition
117
Denote a = ( « i , . . . , a f c ) . Assume that there exist the Lagrange function saddle points correspond ing to problem (17.1) and subproblems (17.2) with the fixed a,. For prob lem (17.1) define the so-called interaction functions25 k
h(x) = '£J91{xj),
i=Tjc
i=i
and represent the conditions for coordination as k
at al = ki(x(a)) kl(x(a)) = = ^2$j(xj(atj)), J29o(xj(aj))^
»* == l l,fc, X
(17.3) (17.3)
i=l
(17.4) (17.4)
A!( Q l ) = A 2 (a 2 ) = ■ • • = A n (a„) AI(QX) (a n ) = A,
where Xiiat) is a solution, and A,( Ql ) ) i s vector rf the eagrange optimal multipliers of the ith subproblem (17.2). (1) Show that system (17.3), (17.4) has a solution, i.e. there exists the vector a = (&u...,ak) satisfying the condition for coordinating (17.3), (17.4). Assume that we know a solution to problem (17.1) x = ( $ i , . . . , 2 t ) and a vector of the Lagrange eptimal multipliers A = (Xu ..j, A m ). In .his case *:
/
k
A >E A
fc
\
k
/
k
A E x
\
6
E /<(*<) + *(**) - M < E *(*«) + ^) E /<(*<) + > E *(**) - M < E *(*«) + [ E ^ ) -& <53/s(*o+ ( * . E #(*<)-&) -
VAt=l > 0, Vxi€ $,i V =i=l l,k. / V A > 0 , Wxi 6 5, i = L X Setting in inequalities (17.5) ^ = £<, i / fc, obtain
fk(x-k)+
(A-E^^
- 6
)
(17.5) (17.5)
- /*(*«)+ (*»]£ »(**)-&)
< &(**) + (A,«(*S) + E&(«i) - 6 ) ' VA * 0, Va;t e s-k . i^k
118
Systems Optimization
Methodology
Select a — (<5i,... ,ak) so that I al = =2 ^29j(x &i j f i (J^),) 5
i=l,k. « = i,fc.
(17.6)
J=I
Considering (17.6), rewrite the above inequalities /s(^fc) + (A,gfc(ij)+Qjt - & } < / 5 ( i j ) + ( A , 3 j ( x j ) + a f c - 6 ) < /*(*j) < h(x-k) + (X,ft<*i) + QJ <*l- -6),b), VA VA> >0,Vx 0,Vx s ss.s . s sGG Thus, we obtain the conditions of the saddle point for subproblem' (17.2) (17-2) with aj. = ak\ therefore
*xki=x=k(a* ik),( % ) .
fek(a ). \*=A =X k (o!s).
(17.7)
Since the index k is arbitrary, then equalities (17.7) hold for k = l,k. As a result k k
aakk =
^ f f ^2gj(xj{aj)), ;(Xj(Qj)), J=I j?k
\k(ak)
= V(aJ),
j,k =
hk,
which required to be proved. (2) It remains to be seen whether it follows from fulfilling coordination conditions (17.3), (17.4) that ( x i ( a i ) , . . . ,xk(ak)) is a solution to problem (17.1).24 We write the conditions of the saddle point for subproblem (17.2) with ctj = &»: /fl1(x (X )) I ) ) + + (\\ ( A J ,g5l(x (a1l())Q 1 ) )++ Q az, I (1Q 1 ( lX t(a
-- 6b)) < < //II((x,(a X , ( Q 11))))
+ (A*(ai), ji(a<(«i)) + &i - b) < fi(xi) fi(xt) J + (A + (A*(«i),fc(*«) (a l ), 5 l (x l ) + «a t.- -6 )6), , VVA A 1' >>00, , Vx, G*i, £st,i i == T7k. 1,A.
(17.8) (17.8)
119
Parametric Decomposition
Considering conditions (17.3), (17.4), from inequalities (17.8) we get
fi(Xi(&i))+
(*'.£#(**(*>))-&)
+ S jj (( :zr jj (( a< j5 )j )) -) & + [[ ^^ ^^ S - &])
^ /
+ (A,ft(xi)-|-ai (A,gi(xi) + on -b), V x t£ € sSj, i=T7k. VA1 > 0 , Vxi = l,fe. t, t
(17.9)
Perform summation of inequalities (17.9) by the index i. Denote k
£ A « = A. ii=\ =l
Owing to arbitrariness and non-negativity of the vectors A1 (i = T7ifc) the vector A is also arbitrary and nonnegative. Obtain fc k
(/
k
V
Jj = = ll
fc
\\
^T ]T] fi(xi(ai)) fi(xi(&i)) + ix^gjixjia^-b] i=l
(
k
// \
\,J29j(X](aj))-b\ X,J29j(xj(aj))-b\ j=\
kk
<^/i(x,(at)) i=l i=l
k
<^2fi(xi) I
i=\
+ (\J2[9l{xi)+al-bU,\/X>0,Vxiesi,i + (\^[9l{xi)+al-bU,\/\>0,Vxiesi,i
= l,k. = T7k.
(17.10) (17.10)
From the saddle point conditions for subproblem (17.2) (with a% = on) it follows that
nn(X
x &9j(xj{&j)) -b) X,^2 A, ^jT g3{xj(aj)) [ ^9j( 3i j)) -b) = = I( *> ]C9j(xj{&j))
1 J=0.
120
Systems Optimization
Methodology
Considering this fact, we get {\,Yi[gl(xi) (
+
a,-b}\
k
k
k
1 3
\
6
A.E&teHEE^ ^ ^-" )/} tt=l =l i=\ i=\ j = \\ = UJ29t(x,)-a\
+ (\,(n-l)
][><*;(*,))-& \
= UJ29t(x,)-a\ + (\,(n-l) = ^A,|><:*MJ
][><*;(**))-& )
Thus, inequalities (17.10) can be represented as k
I
k
t=i k
V I
k
\
Thus, inequalities (17.10) be represented E /.(x,(«,)) + can A, E toteM) - asM .=i k
< E A(**(«0)
/ \
«=i k
£ /*(*<(*)) + A, E fttete)) - M < E A(**(«0)
si
V ti
+
}
(\±9^))-b\<±MXl)
ti
+ ( A , Y > ( » 0 - A VA > 0; Vxt € *, i = ljfe \
.-1
/
(17.11)
,_1
/ are * the conditions \ Inequalities (17.11) for the saddle point— of (17.1), and, + A, V gtixt) b , V A > .,x 0;kVx , i = 1, k a solution (17.11) consequently, the vector x(a) = (*,(Qi),.. (akt))€ sxconstitutes
V x(a) S? = x, whereas ) the vector of the Lagrange optimal to problem (1), i.e. Inequalities (17.11) are the conditions for the saddle (17.1), and, multipliers A common to all subproblems (17.2) is thepoint sameofalso for probconsequently, the that vectorconditions x(a) = (*,(Si),... ,xk(ataken solution lem (17.1). Note (17.3), (17.4), singly, areanot suffik)) constitutes to problem i.e.statements x(a) = x, but whereas the vector are of the Lagrange optimal cient for the(1), above if subproblems chosen as (17.2), then multipliers(17.4) A common to all subproblems (17.2) is theproblem same also for 24 probcondition is the necessary one for coordinating (17.1). lem (17.1). Note that conditions (17.3), (17.4), taken singly, are not sufficient the above statementsinbut subproblems as (17.2), then §17.2.forChoose subproblems theiffollowing way:are forchosen i = I~fK condition (17.4) is the necessary one for coordinating problem (17.1).24 min i /c(x«) + E ^ i . f t i C ^ ) ) f > §17.2. Choose subproblems in the following way: for i = T~fK
(17.12)
9u{xi) + a>i < bit XiG Si,\ > min I fi(xi) + EC&'&iO**)) I
9u{xi)
j
=
l
+ a>i < bit
XiG
)
Si,
/-i j
-i n\
Parametric
121
Decomposition
for i = m + 1,k y2(/3j,9ji{x)) min \I fi(x{) + J2^'9ji(x)) I jTi
I\ , J
(17.13)
JL'i t &x
where a = ( a l t . . . , a r o ) , 0 - ( f t , o . . , A n ) are the coordinating parameter vectors. Introduce the following coordination conditions: m
k
Qi gi2 (Xj{(Xj ,0))+ ^V " 9i3 9i(Xj i = = V ^9ij(xj(o<j,P))+ i(%{13)), i(^))' 3—1 j3=^+i = m-fl i?\ a
0i = \i(ati,l3),
(17.14) (17.14)
i= = l ,l,m, m,
where z t (a,,/J), Ai(a i) /J)(z = I ^ S ) are respectively a solution and a vec tor of the Lagrange optimal multipliers in the tth subproblem (12), and Xi(!3) (* = m + l,fc) is a solution to the tth subproblem (17.13). On the assumption that the saddle points of the Lagrange functions of problems (17.1), (17.12) - (17.13) exist we have proved that the statements similar to statements (1) and (2), §17.1. 38 §17.3. Consider the following expansion of problem (17.1) for i = T~^ min \I fi(xi) ^2(Pj,9jz{x fi(Xi) + Y^{P ,Ui) r)) l) I\ J,g^{x l)) - (0tt,u
9»(a;i) + M i < 6 i ,
*<€»*,
(17.15)
(17.16)
for * = m + 1,k min I^ fi(xi) /i(^i) + + ^2(/3j,9ji(xi)) 5j(&,9ji(»*)) f>,> xXi ; €€ ss*{ where ui,...,um are the connecting variables, (3 = ( f t , . . . ,/3 m ) is the coordinating parameter vector. In this case the following theorems 37 are valid. Theorem 17.1: Assume that there exists the vector {3 = ( f t , . . . , / 3 m ) such that the solutions («<,«,-),» = T~^, X, i( = m + l,k) of subproblems
122
Systems Optimization
Methodology
(17.15), (17.16) satisfy the coordination conditions k
Ui = ^29IJ(XJ),
(i -
M") •
i=i
Then x = (xi,...,xn), 0 = (0i,--.,0m) conditions for subproblem (17.1)
satisfy the following saddle point
(1) x minimizes the function
X>(^)+ £/<(*<) + (&][>(»*)-&) (^£>(**)-M \
i=l
i=l
/
on the set Si x • ■ ■ • xt; (2) 0 > 0; (2) 0 > 0; k k
(3) (3) £>(*<)<&, £>(xi)<&, (4) (ft ] [ > ( * . ) - M =0. =o. Thus, from theorem 17.1 it follows that x is a solution to problem (17.1). T h e o r e m 17.2: Assume that x = (xlt... ,xk), 0 = (ft,... ,0m) satis fies conditions (1) - (4) of theorem 17.1, i.e. (£,/?) is the saddle point for problem (17.1) and k
"t = "t = 7Jgij(%),
(i (i = = 1,TO). l,m).
Then (£*,«*) (i = I~m) and & (« = m + l,fc) constitute solutions to proper subproblems (17.15), (17.16) with 0 = 0. The theorem implies that the vector 0 = (0U ... ,0m) described in theo rem (17.1) exists. From theorem (17.2) it follows that such a decomposition technique for solving problem (17.1) is possible only when the saddle point of problem (17.1) exists.
Parametric
Decomposition
123
In conclusion, it might be noted that in decomposition (17.2) the global constraints of problem (17.1) are expanded into components in a "vertical" way, whereas in decompositions (17.12), (17.13), (17.15), (17.16) expansions are implemented in a "horizontally diagonal" way. §18. ONE M E T H O D OF P A R A M E T R I C D E C O M P O S I T I O N FOR L I N E A R A N D SEPARABLE P R O G R A M M I N G PROBLEMS §18.1. The present paragraph deals with the method of parametric decom position 12 associated with introducing the parameters that enable one to partition at one's discretion the original problem into independent blocks. It is wise to employ the method when the constraint matrices possess a large number of zero elements, although not having a marked block structure. We will set forth the main idea of the method using the following example. It is required to find min(c, x) = (ciXi + ■ ■ ■ + crxr) + (cr+iXT+i
+
h cnxn),
(aiXi + ■ ■ ■ + arxT) + ( a r + i x r + 1 + • • • + anxn) X \ , - ■■ ■ -. .1- ) ,
Introduce the variables yuy2
-^
< b,
\J ■
so that
ixr+r+ii a\X\ + ■ ■ ■ + aTxr < yi; ar+ r+ix
+ + ■ • •■• ■+ +anaxnxn <
2/i+2/2 <
Tl
a x
V\ = Z^ J j j=l j=l
> V2 = 2 ^ aixi i—r+l i—r+l
'
Vi +V2 ,
124
Systems
Optimization
Methodology
therefore T{y**) > T{y"), i.e. (c,x*) < (c,x**). Thus, the LP original problem reduces so finding the F(yy minimum and solving tht problems of smaller dimensions. §18.2. Consider now the general case. Find n
min(c, x) = ^ J CiXi, c&i, j=i
*n 2 CLijXj < bi,
. i=
i = 1, 771 ,
(18.1)
2 o-ijXj < bi, l,m, i=i x-j = ll ,, nn .. Xj > > 0, 0, jj — 3=1
Represent
n 7-j—11 n 'fc+l >fc+l n—
n 3j ==l1
k=0 j=nikk k=0 3=n<
and rewrite problem (18.1) in the form (18.2) (18.2)
min(c,x), min(c,x), yJ2^ ^jXjKy^, OjjXj < yik,
fc=l,...,T„ fc=l,...,r„
j=nik
Xj>0, Xj > 0,
j = hn, l,n,
ii = = V l, m m,,
(18.3) (18.3) (18.4)
where the parameters ylk are such that Y,ykik
i = T^i. T~m.
(18.5)
k=\ k=i
Problems (18.2) - (18.4) are called the subproblems of problem (18.1). The optimal value of the objective function in (18.2) - (18.4) is the function from y = {yik, k = I~F~, i = T~rH} and we denote it by F{y) with the domain of definition Y. We set forth the problem equivalent to (18.2) (18.4): m inJF^), mmT{y), r, (18.6) ST su bi, i• =T~m. i (18.6) ^Tyik < v 2_^yi
Parametric
Decomposition
125
Let y' be a solution to (18.6), x* a solution to the subproblem with y = »*, x** a solution to the original problem. Theorem 18.1: T(y*) = (c,x*) = {c,x**). The proof of the theorem fully repeats the proof of the corresponding statement in §18.1. §18.3. Thus, we have reduced the solving of the original problem to solv ing a nonlinear equivalent problem (18.6). The functions of the form T(y) are convex, but not smooth, therefore, for T{y) minimization it is possible to employ, say, the method of subgradient projection.4 Let F(y) = i^ihiv)}' fc = l i r i ' * = l>m> D is the feasible region of problem (18.6).
K{y) = Kiv), •£*(») K(y),
(18.7) (18-7)
where \ik(y) are the dual variables complying with an optimal solution to subproblem (18.2) - (18.4) (with the specified y). Then, according to the subgradient projection method, a search for the solution y* of the equivalent problem and its corresponding solution of the subproblem amounts to the following procedure. Let y° = {y° } be an arbitrary point of the region D, y" = the point obtainedafter the sth step, x' the solution of (18.2) - (18.4) with y = y", Xs = {\lk{ys)} are the corresponding dual variables. Then ys+1
= p(ys -
s Ps\(y )),
5 = 0,1,...,
where p is the operation of projecting onto the region D, ps the step descent value which, for the convergence of the sequence {ys}y* is chosen so that Ps>0,Epa
= oo, e.g. ps = (* + I ) " 1 , s = 0 , 1 , . . . . The subgradient
3 =0
projection method does not ensure fulfillment of the monotonicity condition Hys) > W+1) at each iteration and, in general, slowly converges, but if the points on the set D and the subgradient T are found not in a complex way, then the method is rather simple for computer implementation. Note that, in practice, the T{y) function definition domain is most commonly specified by the form aaiikh < yik < blk . In this case, projection onto the region D reduces to simple calculations.
126
Systems Optimization
Methodology
§18.4. Consider the application of the above parametric decomposition approach to the separable programming problem: n
min^2 min ^2 C JCjixj), ixi) > i=l 3=1
n
^2a,ij(xj) > 0, j = = l ,l,n. n. j=i
The subproblem of form (18.2) - (18.4) in this case takes the form n
m i n^2^ Cj-(xj;), min cj(xj), J=I
aatyj{xj) fo) < 2/*i, Ji = m;; jj = = 1,l,n; n;
"YLvij X^y <-bbi,"
ii ==1 7n ' l,m. -
3= 1
In the present case, the subproblem reduces to the simplest problems of nonlinear programming. For the separable programming problems, the equivalent problem is solved by the subgradient projection method when Cj(x), aij(x) are convex functions.
Chapter 5 DECOMPOSITION BASED ON SEPARATION OF V A R I A B L E S
?19. C O N S T R A I N T RELAXATION M E T H O D FOR T H E C O N V E X P R O G R A M M I N G PROBLEM §19.1. Consider the convex programming problem
(19.1)
where /,gi (i — l , m ) are the convex functions, 5 is the convex subse Consider the case with the number of constraints being rather large. Then the reasonable strategy of solution consists in relaxation, i.e. tempo rary rejection of some of these constraints and solutions of the remaining problem. If the problem is unsolvable, then so is the original problem. If it is solvable, then provided the solution obtained satisfies the relaxed constraints, it is also optimal for original problem (19.1). Otherwise, it is necessary to introduce one or more of the constraints omitted and repeat the procedure. The relaxation method is particularly effective when it is known in advance that a small number of constraints is substantial at the optimal point.
Denote M = {1,2,..., m} and let R C M. Introduce into consideration the problem 127
Systems Optimization Methodology
128
max f(x) ,
9i(x)>0, ieR, (19.2) x € s.
§19.2. We will set forth the relaxation method algorithm. Assume that it is possible to find such an initial set R that (19.2) has a finite maximum. Step 0: Suppose / = oo and choose the initial set R such that / is bounded on the constraints of problem (19.2). Step 1: Solve (19.2). If it is unsolvable, then (19.1) has no solution.
Otherwise we get the solution xR f of problem (19.2).
Step 2: If gi(xR), i € M \ R, xR is optimal for original problem (19.1). Go to step 6.
Step 3: Otherwise let V be the subset M \R involving the indices of, at least, one of the unfulfilled constraints. Introduce the subset of indices D C R, for which the constraints gt(x) > 0 are fulfilled
strictly on the optimal solution xR:
D = {i\9i(xR) > 0, i e R} ? = /, then replace R by R' = R U V. Go to step 1. Step 4: If f(xR) = ) Step 5: If f(xR) < /, then replace R by R' = R U V \ D set /f(= f{xR). Go to step 1. Step 6: Terminate. Thus, if the functional optimal value in (19.2) decreases, then the vio
lated constraints are introduced and the constraints inessential for xR are removed. If the optimum is maintained, only then is introduction of con straints performed. §19.3. Show that the above method, after solving a finite number of prob lems (19.2) either leads to the optimal solution of original problem (19.1) or the unsolvability of the problem is established. Call problem (19.2) the p fl -problem. ????????????????????????????????????????????????????????????????????????
Further, if xR is a unique solution to pR, then it a unique solution to pR~D as well.
Decomposition
Based on Separation
of Variables
129
Proof: Suppose we have the vector x satisfying the constraints {R \ D}
and such that f(x) > f(xR). C Consider the points
xx = Xi + (1 - X)xR, 0
9i(xx) > X9i(x) + (1 - X) 9i(xR). Since gi{xR) ) > 0, i 6 Dt then for the sufficiently small Xgl(x\) > 0, i £ D.
Therefore, for such A the point xj is feasible in pR. However, due to the concavity of /
f(xx)>Xf(x) + (l-X)f(xR)>f(xR), which contradicts the optimality of xR. Theorem 19.2:
}{xR')
Proof: If R' — R U V, then the theorem evidently holds. Due to theo rem (19.1):
f(xR\D) f= f(xR),
hence
?????????????????????? Theorem 19.3: Relaxation procedure terminates after having solved a
finite number of the problems PR and leads either to the optimal solution of the original problem or to the construction of such constraints set that the current problem PR proves to be unsolvable. Proof: Since the set M is finite, then it has a finite number of different subsets. With f(x) decreasing from iteration to iteration, none of the sub sets can be repeated. Inasmuch as none of the constraints is withdrawn, if
f(xR) = f(xR), and at least one constraint is added, then / can remain
130
Systems Optimization
Methodology
constant only over a finite number of iterations. Thus, for the finite number of steps the procedure must terminate at step 1 or step 2. §20. T H E R I T T E R M E T H O D F O R T H E B L O C K P R O B L E M W I T H T H E TYING VARIABLES AND CONSTRAINTS §20.1. Consider the LP block problem with the tying variables and constraints: p
max z2 = ^2ciXi+c 2_] ctXi +0y,coy,
(20.1) (20.1)
11 = = 11 p
^Y,2 AiXi Mxi + D00y = b00 ,
(20.2) (20.2)
i = il
B D1y = bbt, BiXi ixl + Djit %, xxz > 0,
i= - T7p, l,p,
i ==T ~l ,pp
(20.3)
y>0 y >0
(20.4)
and the problem dual with respect to it: p
min i9 btUi ++ M bouo 0 = yJ2j fcitH o ,,
(20.5)
i=l V P
Y] ^2 D[u D\uxx + D'0u0 > c0 ,
(20.6)
~{ i=l B'iUi + AiU > c,, ct, B\ui + AiU0 0 >
= l,p, i=
where Ai, B,, 13, — (mo x m), (m-i x n{), (m, x no) are matrices, respectively. Ritter 32 has developed the relaxation method for the problems of form (19.1). At each iteration, variable problems separate into two subsets - the subset Si, in which non-negativity constraints are relaxed, and the subset S2, in which they are retained. If the variables in Si prove to be nonnegative, then the current solution will be optimal for the original problem. Assume that each matrix Bt (i = T~p) has the rank ra< and, conse quently, contains a nonsingular submatrix Bh. Suppose the initial vector yo is known such that the feasible solution of the constraint system exists BiXi = b% - Diy0,
i=
l,p,
Decomposition
Based on Separation
of Variables
131
then a set of initial bases can be found by solving subproblems max c^Xi, BiXi
= bi-
Diy0,
Xi>0
and using optimal basis matrices as Bn. (20.3) can be rewritten as BhxH
After isolating Bh
constraints
(20.7)
+ Bi2%2 +B xi2x%2 +D+D = b tl. ty ly=^b
Since each matrix Bh is nonsingular, equation (20.7) can be solved with respect to X{1: xh = BrHi - B~lBl2xl2
- B-'D.y
.
(20.8)
Divide variables in objective function (20.1) and tying constraint (20.2): p
z= = ^{c ^ ( ncx^n x ^ +c12 t2)) l2xxi2
+c+c 0y,0y,
(20.9)
t=i i=i p
Y^{A ^(A llxnhx,x
+ Al2i2xi2)
+ D0y = b0 .
(20.10)
i=i
Substituting (20.8) in (20.9), (20.10), obtain the reduced problem max < z — a = ^ J diii2 + d0y > , pP
^2 2^ M Mttxxl2l2 + +M M00yy = = b, b, i=i i=i
Xt2>0,
y>0,
where p
a =
~^2,cnBnlbi^
i=l
&i — Ci2 — Ci2tS^^ t>i2 ,
(20.11)
132
Systems
Optimization P p
Methodology
^2c ^2cnnB~ B~11DDzz,,
d0 ==cc00-do i=l
Ai2-AilB~xBi3,
Mi =
p
Mo =
D0-Y,AnBr'D>, i=l
A ilBB-b1bi. bb = 0-Y,A = b0b-J2 n n >tt=i =l
The problem has as many constraints as the original problem contains the tying constraints. The constraints in (20.11) prescribe the same set of feasible solutions as initial constraints (20.2) - (20.3) do. However, since the reduced problem does not explicitly involve the variables x M , then it is impossible tofixin it that xn > 0, i.e. the constraints relax. The feasible set of problems (20.11) evidently involves a feasible set of the original problem so that if the reduced problem is unsolvable, then so is the original. It is evident, however, that the reduced problem can have an unconstrained solution in those cases when the original problem has a finite optimal solution. To exclude the possibility, one can add into the reduced problem the constraint of the form pp
J2 ^ 2. +l'oV
(20.12) (20.12)
1i=i =1
where t is a sufficiently large positive number. If, on completion of algorithm operation, this constraint proves to be significant, then the optimum is unbounded. Suppose that an optimal solution to the reduced problem {{x°l2},y0) has been found. Substituting it in the right-hand side of (20.8), we obtain the solution ( { i ^ } , {Xi2},Vo) of the original problem without assuming x^ > 0. If the constraints are fulfilled, d,e solution ontained is the solution of the original problem. T h e o r e m 20.1: (Optimality criterion). The solution ({x°J, {i^},y 0 ) constitutes the solution of the original problem if < < >> 00 ,,
1, p. i2 = = hp.
(20.13)
§20.2. Consider the reduced problem transformation in the case of unfulfillment of optimaJity condition (20.13). In this case it is necessary to
Decomposition
Based on Separation
of Variables
133
construct a new reduced problem, in which some of the violated conditions for non-negativity are introduced as necessary, whereas the requirements for non-negativity of the varrables si3, for which i ° x ° >, are emitted. Consider an arbitrary vector xh characterized by the fact that even one of the components se§, in the solution obtained is negative. For definiteness, assume that the first / of the components x°h are negative: (x°n )i 9) 2<<0,...,(4), 0 , . . . , {x°h )i < 0, // > «(x°)n i)i< <0 0,, (x° <0, > 11
(20.14)
Suppose x°2 has the first q components being positive: ([x%)i * £ ) , > 00,, {x° (x°j2J2))22>0,...,(x° > 0 , . . . ,n){x° q>0. > 0. q>Q, n)q > 0, q Let us write the relation of (20.8) connecting xh,xn xh + i2xn */. + BBj*B nBn*n
1
(20.15)
as:
Bj^Djy. = ^&"^ jbj -- B^D . 0
(20.16) (20.16)
Two cases are possible here. Case 20.1: There exists at least one nonzero element of the submatrix formed by the first / rows and the first q columns in Bj*Bh. In this case, one can perform the leading operation using, as the leading element, an arbitrary nonzero element of the submatrix mentioned. Such an operation transforms (20.16), introducing into the basis the component xn positive in x°h and withdrawing the component xn negative in x^ This substitution leads to a new basic matrix Bh which is employed for constructing a new reduced problem. Case 20.2: The submatrix formed by the first / rows and the first q columns B~lBj2 contains only zero elements. Then it is impossible to accomplish substitution of the basis. Choose a negative component x%x,ay, (a£ )k)* 1 k < I and use (16) to express explicitly the condition for non-negativity of this variable by xh and y.
(x n)k=a 3ky>0,~dJky>0, (*Ah =jk-b a»Jkx-n-dK*h where ajk, bJk, dJk are the Jbth rows of the matrices
Br%,B^Bhh,BJ Br%B^B ,BJ1illD Djj
(20.17) (20.17)
134
Systems
Optimization
Methodology
respectively. This constraint is added to the cases 1 or 2 for each vector i°n possessing negative components. We come to a new reduced problem in which the matrices Bh can be changed and the constraints of form (20.17) can be added. Before applying the procedures indicated in cases 1 and 2, these constraints are to be examined according to the rules given below. Case 20.3: If constraints (20.17) are not significant on the optimal solu tion, i.e. if a
n -hu -huxxhh
~dj ~djkty° y°
> o0 ,
then the constraints are withdrawn from the reduced problem. By virtue of theorem (9.1), §9 this does not alter the optimal solution of the problem as a whole. Case 20.4: If constraints (20.17) are significant on the optimal solution, i.e. if (20.18) a<*n n-bnx° ]2-dny°=0, -bj*x°h ~djky° = ° > then the constraint is omitted when alteration is possible in the basis related to substituting (xn)k by a component xJ2 which is positive in z%. Here are visible the elements lik related to the positive components x°2; if there is a nonzero element, then it is used as the leading and, after performing the leading operation, an appropriate constraint of (20.17) is omitted, and the optimal solution again remains unaltered. Case 20.5: If (20.18) is fulfilled and all the components bh corresponding to the positive components in x°n are equal to zero, then constraints (20.17) are preserved and conditions (20.18) take the form
dJk]ky°=a y°=aJk3k §20.3. Formulate the algorithm for solving problem (20.1) - (20.4). Step 1: Choose the initial basic matrices Bil(i=T^p). This can be done in an arbitrary way or by solving the subproblems max BiXi =bi-
aii,
Diy0,
xl>0
Decomposition
Based on Separation of Variables
135
and using their opttmal lases as Bn. Reduced subproblem (20.21) is constructed with the use of these matrices. Step 2: Obtain ({x°2},y°) as a solution to (20.11). Calculate ar? from formula (20.8). If x\ > 0, % = T ^ , ,hen ( ( < } , « } , 2 / ° ) ii sa optimal solution to original problem (20.1) - (20.4). Go to step 6. Step 3: If x^ t 0, i 6 / and the additional constraints of form (20.17) exist in the reduced problem, then we consider them with the use of the procedures indicated in cases 3-5. Step 4' The leading operation is used for each i <E / , as in case 1, or constraints are added, as in case 2. Step 5: Using the new basic matrices and constraints obtained at stages (steps) 3-4, we construct a new reduced problem and go to step 2. Step 6: Terminate. §20.4. Consider the issue of convergence. Since the above algorithm constitutes a relaxation procedure, then theo rems 9.1, §9 can be applied. Owing to theorem 9.2, §9, the sequence of objective functions obtained when optimizing the reduced problems, is monotonic decreasing. Theorem 9.3, §9 yields the following conditions for convergence for the finite number of iterations. Theorem 20.2: If the requirements for non-negativity are omitted only on the iterations when the reduced problem objective function decreases, then the algorithm converges, for the finite number of iterations, either to the optimal solution of original problem (20.1) - (20.4) or to constructing the reduced problem without a feasible solution. §20.5. Show that the Ritter algorithm is dual. Theorem 20.3: Each solution ( « } , {x°2},y°) obtained on the fcth ite ration corresponds to an extreme point of the feasible set of dual problem (20.5), (20.6) {u°t} with equal objective functions: v V
P
£ MM?° == cc00y°y°++£ 5>>ni <, < + + cci2l2x°x° J2 l2). l2). i=\
i=l
(20.19) (20.19)
136
Systems
Optimization
Methodology
Proof: The solution ({x°it}, « } , y°) is the solution of the modified origi nal problem prescribed by conditions (20.1) - (20.3) and requirements for non-negativity: zi2>0,
i=T~p,
y>0,
{xiki)ik>0,
(20.20) (20.21)
where (20.21) corresponds to the fcth constraint of form (20.17). The prob lem dual with respect to the modified original one involves the equality instead of inequalities (20.6) corresponding to the components {x{l} not restricted in sign. By the duality theorem, if the modified original prob lem is a finite optimal solution, then the problem dual to it also has this solution. Therefore, the modified dual problem achieves an optimum at the extreme point {u°} satisfying (20.19). And since any extreme point of the modified dual problem simultaneously represents the extreme point of original dual problem (20.5), (20.6), then the proof of the theorem is completed. §21. T H E ROSEN DIVISION M E T H O D FOR LINEAR P R O G R A M M I N G PROBLEMS §21.1. Rosen proposed the division method for solving the pair of dual problems of the form: Direct Problem: m i n\zL = min = ^2_.c cliXix I l\ ,'
(21.1)
p
^/ , ^Aix^i
60,, == &o
(21.2)
i=l
BiXi bi, BiXi = h,
x, *. > 00,,
i ==T 7 l,p, p, iZ == T7P, l,p,
(21.3) (21.4) (21.4)
Dual Problem: max < lvv = ^^ Y/biubiU y\,b0y > , x 0+ l+b
(21.5)
£>. + Blm + A'iit 4 2 / <
(21.6)
ii = = T~p~, l,p,
Decomposition
Baaed on Separation
of Variables
137
where the matrices Ait Bi are the (mo x m ) , (m, xn,)-dimensional matrices, respectively. The Rosen method can be regarded as a specialization of the Ritter method described in the preceding paragraph. We show here that a series of simplifications is possible as opposed to the Ritter method: (1) it is not required to introduce additional constraints into a reduced problem so that the problem will have no more than mo rows; (2) consecutive reductions of the problem can be solved by the dual simplex method. Further in this paragraph, we will employ the following assumption. Assumption 21.1: The direct problem constraint matrix has the full rank. Theorem 21.1: If assumption 1 is fulfilled, then each matrix Bi contains the nonsingular submatrix Bn of the m, dimension. Proof: If Bi has the rank smaller than mu then rows are of the form (21.3), which contradicts assumption (21.1). Therefore, m > muBi contains the nonsingular rubmatrix Bn. We cacalsa assume that nt > m, sincsinc nl=mt, then the variables x{ are uniquely defined. The initial set of basic matrices can be obtained by solving the problems: m i n CiXii C{Xii
BiXi=bi,
xx{>0. ;>0.
(21.7)
Assumption 21.2: The ititial basic matrices Bn satisfy thc conditions
B-%^0, B-%^0,
i i=T?p. = l,p.
The basic matrices B^ are employed for transforming direct and dual problems: Direct Problem: m i n Y^( ^ ( ccihl xaii: i 1 + a2xal22),xi2), min
(21.8)
138
Systems Optimization Methodology
2^(Ai / . ( ^1iXi i ^1' i + "I" A{ A%2tXi Xi22)) =— "t> 0oJ i B 1X{^1 fJi
+ ~rB±Jii2^i2^ ^ —; &i ,
x
«ij n > 0,
xi2i2 > 0,
t
(21.9)
i4 == l,p, l,p,
Dual Problem: m aaxx^^( 6( iM U ,2 + b &00y), y),
(21.10)
i
B'nuz i+ + A'ixy i 2 /
S'*
;
-
B[2uz + A'l2y< cl2
'
(21.11)
As in the Ritter method, relaxing the constraints xzi > 0 and using (21.9) for exclusion of xh from the coupling constraints of (21.8), we come to a reduced problem. Thus 1 l l xtln = B~lbt =B- B~ bl-B~ Bl2x, B2t2xl2.
(21.12)
The reduced problem takes the form: min 2y J diXi2 , y
^Mixl2
= b0,
xt2 > 0,
(21.13)
where d -ciXljBJ HBlB"l21 ^ , , , d,%==c,c2l2- C Ait-AilB-lBia,
Mi =
l bbo 0 == bb Q-J2Ar t. -J2^1BB-n1b^-
Q
1i=l =1
Evidently, the feasible set determined by (21.12), (21.13) comprises the original problem feasible set so that if the reduced problem has no feasible solutions, then the original problem is unsolvable. The reduced problem can have an unconstrained solution whereas the original problem has a finite optimal solution, therefore, as above, it is possible to employ the constraint of the form /
. *'jXh
—* !
where t is a sufficiently large positive number.
Decomposition
Based on Separation
of Variables
139
Let {x°J be the solution of (21.13). By formula (21.12), obtain the values of a*,:
xl^B-lbi-B^Bhx%,
z = l,p.
Theorem 21.2: (optimality criteria). The vectors { < } , { < } constitute a solution to the original problem if and only if p. i ==L 7l,p.
<>0,
(21.14)
If the optimality criteria is not satisfied, then, as above, it is necessary to take account of at least one of the unviolated constraints of non-negativity and, in doing so, relax the constraints of non-negativity on positive vari ables. In the Ritter method, for the problem with the tying variables and the tying constraints, we did not always succeed in performing this trans formation with the aid of the leading operation and we had to introduce additional constraints. In this case, however, the sequence of the leading operations, which leads to a solution, always exists. Theorem 21.3 3 3 : If, for an i, x° has negative components, then we can form a sequence of the leading operations in the ith block introducing into the basis the component xl2, that has been positive in x° 2 , and the compo nent xh, that has been negative in x°h. The proof of the theorem needs the following lemma. L e m m a 21.1 3 9: Consider the linear system Ax = b, where A is the (mxn) matrix, n > m, rank (A) = m. If the system has a non-negative solution, then it also has a basic non-negative solution. The proof of theorem (21.3). For notation simplicity, omit the i block index and consider a particular block. There are two solutions of the system BiXi =b,b, Bixi + B2x2 =
ixii >>00,,
x22 >> 00,,
(21.15)
such as the feasible e[xjj0]'] ,he unfeasible [x[,x°], where xBl = X~lb. The former solution is non-negative (see assumption 21.2) while the latter has negative components in x?. All the vectors of the form
+e \ +0 <( 1 i- -0 e) to'] [:[ L o j [x XBl
X
2
(21.16) (21.16)
140
Systems Optimization
Methodology
BiXi + BiXi + B B22xx22
x2 >x2>0. 0.
satisfy the conditions = = b, b,
Let / = {j■ -. z°. < 0}. The largest 0 , with which (21.16) remains nonnegative, is bounded by the components with the indices j € /. Under the assumption that these components in (21.16) are equal to null, obtain
xBlBx.. + e(x00hh - xxBBll ,), ) = o, 0, j e £ Ii or
xxBl Bl
0> &j = =
i*i*11--
Sr> S r>
xBlj - x\. Note that 0 < 9 , < 1. The largest 0 , with which (21.16) remains negative, is equal Q Qss = =m m ii nn 00 ,, .. Let K = {k:x°2h 2k>
0}
must be nonempty since if all x%h = 0, then x\ = B^b > 0, which contra dicts our hypothesis that x°x has negative components. When 0 = 0 S the vector (21.16) satisfies the equation Bl Bi(xBl
+e03(x° = 6b,1-xBl + - ))+ xBl)) + 5Y^B$(O 2 * a (sx° 9 «2k*) S j = 3 (x°
(21.17)
k&K
where B2 is the fcth column B2. By defning 0,s the coefficient with hhe column B{ is equal to null. Since the matrix Bx is unexpressed and has (m x m) dimension, the matrix [BUB2\ has the rank m. Omit the column B{ in (17) and check the rank of the system obtained. Multiplying (21.17) by B~\ obtain m
2(®sx° ) == B^b, 52^ + k€K 52^(0,0 v> i=l £ » & + S 2=1
S
2k
(21.18)
fce/c
where v° is the ith component (xBl + ®3(i? - xBl)), #2 the fcth column Bi1B2. The system has the same rank as system (21.17), without Bf.
Decomposition
Based on Separation of Variables
Moreover, at least one of the vectors B%,k £ K say, Bf nonzero sth component because for j £/ the vector
141
must have a
-B~ -B-1BxB2x° 2xl2
x ? =Blz B l x\=x
has negative components. Since s £ I, if all the components B f 1 5 2 in the row 5 and columns k £ K are equal to zero, then x
As ls
X =XB = BISIS
>> 00 ,,
which yields a conttadiction. Thuss,he vectors lt, i = T~^, , ± s, B\' are independent, and system (21.17) with the omitted B{ has the rank m. Since the system has a non-negative solution, then, due to lemma 21.1, it also has a basic non-negative solution. Any solution of that kind does not comprise B{ and can include only the columns B%,k€ K. The proof is completed. The theorem does not state that the unique leading operation, giving rise to a new basis £ „ exists. These bases, however, can be obtained from solving the problem: min xis xi3 , 771
j ^ B f x i f c + Y,B2x2k k=l
= b,
(21.19)
keK
xik xik > > 0, 0,
x2k x2k >>0, 0,
where s is the index selected from the condition 00 S ==mmin i n 0 Qj. , . s jei 361
The initial basis in the problem is Bx. solved for each non-negative block.
3
(21.20) (21.20) v
That kind of problem should be
§21.2. Formulate the algorithm for solving problem (21.1) - (21.4). Step 1: Choose the initial bases Bx for each block. It can be done by solving subproblem (21.7).
142
Systems
Optimization
Methodology
Step 2: With the use of the current set of bases, construct and solve the reduced problem (21.13), obtaining optimal solutions {x%}. Step 3: Calculate x° from formula (21.12). Step 4- If j ° > 0, i = hp, then {<,<}
is a solution to the original
problem (21.1)-(21.4). Go to step 7. Step 5: Otherwise for each i such that x°n ^ 0 we calculate the index s according to (21.20). Solve problem (21.19). Step 6: Using the new basic matrices Bn obtained when solving (21.19), go to step 2. Step 7: Terminate. §21.3. A few words must be said about the above algorithm convergence. Since the algorithm constitutes special implementation of the relaxation procedure, convergence for the finite number of iterations follows from the orem (10.2), §10. In this case, however, we need not take account of the non-negativity condition violations by incorporating additional constraints into the reduced problem. If none of the additional constraints is intro duced, then the assumption that each reduced problem has no other optimal solutions satisfying the newly incorporated conditions for non-negativity ensures convergence for the finite number of steps. §22. THE ROSEN DIVISION METHOD FOR NONLINEAR P R O G R A M M I N G §22.1. The Rosen method for LP problems, described in the preceding paragraph, was extended by Rosen to particular nonlinear problems of the form: min I| ^2c ] T i(y)x a(y)x c0(y)\, 1, l r + c0{y) Ai(y)xix>b>l{y), Ai{y)x bi(y),
i=T^, i = l,p,
p>\. p>l.
(22.1) (22.2)
The distinctive feature of the problem structure is that for the fixed y the problem separates into p independent linear subproblems with respect to the variables x t . Thus, the natural iterative procedure exists, in which first y changes, and then xt. The mathematical proof of convergence to
Decomposition
Based on Separation
of Variables
the global minimum mrevails when in (22.(), (22.2) a.Ai(i depend on y, i.e. for the problem of the form
143
A T7p) dp not
p
min 22_,J CiX CiX i z>, i»== li
A tx% > bi(y), AiXi>bi(y),
(22.3)
ii=T~p. = I,p.
(22.4) (22.4)
Consider first this type of problems and then a more general problem (22.1), (22.2). We make additional assumptions: Assumption 22.1: For i = T ^ each fc(y) is the differentiable convex function y. Hence it follows that (22.3), (22.4) constitute the convex pro gramming problem for which the local minimum is global. Assumption 22.2: There exist the feasible vectors (xity), being interior with respect to nonlinear constraints, i.e. such that AiXi > b%(y) where the strict inequality for the nonlinear components b,(y) is fulfilled. This secures existence of the saddle point of the Lagrange function. Assumptions 1, 2 provide that the Kuhn-Tucker conditions are necessary and sufficient for an optimal solution. Assumption 22.3: None of the matrix rows A, (i = Tj>) is a zero vector. Assumption 22.4: The feasible vector y0 exists and is known, i.e. such y0 is known for which the x{ satisfying (22.4) exist. Problem (22.3), (22.4) can be solved via minimization by xt with the ffxed y, and then via minimization of a result by y. min < I V^ Y min{ciXi\AiXi xmniaxilAiXi > bi(y)} \>, ,
yes
where
p v SS = P = Pi| Si, Si,
i=i i=i
J
Ytl
Si Si = = {y\3x {y\3xrr :: AiXi AiXi > > bbrr{y)} {y)} ..
The vectors y G S are termed feasible.
(22.5)
144
Systems Optimization
Methodology
Define the function ipi(y) = min{ciXi\AiXi V>»(j/) = min{c > b&,(?/)} t(y)} .• 1arj|A,i< > Then problem (22.5) can be rewritten as: p
^2iH(y), min i>(y) ^(3/) ==$J^»(y).
(22.6) (22.6)
i= l
jyes. €5.
(22.7)
We will show that under assumption 22.1, problem (22.6), (22.7) is the convex programming problem. Theorem 22.1: If each function bi{y) is convex, then the set 5 is convex and each function V'i iisonvex on 5. Proof: (1) Show convexity of the set S. To do this, it suffices to say that each Si (i = TTrt is convex, i.e. from the fact that Vl € su y2 6 S it follows that Aj/i + (1 - X)y2 € Si, 0 < A < 1. Since yx e s{ 3xt, which ^ ^ > bi(yx). Similarly, 3 xt which 4 & > fcfoa). These inequalities and convexity h(y) imply: k(Alh + (1 - AJjfe) A)jfc) < )6,( y i ) + (1 - A)6i(2/2) bi(Xm + < Ayl.ii XAtXi + + (1 - A)Ai*i A)A
i>l{y2) = ^1(2/2) = minjciilAiXi min{ctXi\AiXi > > 6i(y bi(y22)} )} = = CiXi(y CjXifa). 2).
(22.8) (22.8)
Decomposition
Baaed on Separation of Variables
145
The convexity of k(y) implies bi(X XMy2) bi(Xm A6t(j/!) yi + (1 - A)j/2) < Xbi( yi) + (1 - X)bi(y < Ai(Xxi{yi) Ai(\xi{yi) + (1 - X)xi(y2)), i.e. inequality (22.8) can be rewritten as Ci(Axi(i/i) + (1 - X)xly2) = Xaxi(yi) + (1 - X)xi(y2) ■ The proof of the theorem is completed. The original problem, due to the theorem, reduces to solving the convex programming problem with respect to y. The problem, however, is made difficult by the fact that in order to estimate the function ^(y) we need to solve the linear subproblem of the form min ax,, ax,,
(22.9)
AlXl > bi(y). A{Xi bi(y).
(22.10)
Let yo be a feasible vector. Suppose that each of the problems (22.9), (22.10) can be solved for y = y0- With the aid of this solution we divide the matrices Au which have more rows than columns, into the square nonsingular basic matrices Ah, and the nonbasic matrices Aia. The matrices Ah define the solution vectors x°t by the equations Ai1Xi
= t>i1 . —
Subproblems (22.9), (22.10) are written as: min dXi, c,x,,
(22.11)
xi = hlh(y), AilhXi 1{y),
(22.12)
At2l2xt
(22.13)
> >b bi2{y). i2(y).
We handle (22.12) with respect to xt: x%z %== A-Hi^y). Aii1bil(y).
(22.14) (22.14)
Employ (22.14) to eliminate s< from (22.11), (22.13): *Xi = CiA'%, dXi ciAi1bil{y) (y) = «(u°) b h)b(y), h(y),
(22.15)
AijA^bi^^b^y), Al2A^bH{y)>bl2{y),
(22.16) (22.16)
»« = 1 ,^p ,
146
Systems Optimization
Methodology
where 1 u°ti = C.A' < CiA-1 > 0
(22.17)
is the vector of the dual variables related to the Ak optimal basis. Define the functions ipz{y) by relation (22.15). Here (22.16) defines the set s;. Thus, the coordinating problem in the y space has the objective function p
b (y) f(y) ii)b11 f(y)==^2(u° J2«) n(y)
(22.18)
i=l
and constraints (22.16). Objective function (22.18) is convex since bn are convex and «°n > 0, although constraints (22.16) are nonlinear. This means that the global optimal solution cannot be ensured. Any coordinating pro gram, however, serves only to improve the performance, if any, of the y feasible vector. With this in mind we will linearize constraints (22.16) with respect to the current point y0. Let 1
Qt — = Ai^AQi Ai2A{i
(22.19)
Introduce into consideration Jacobian Ji3{y) where Jl3{y), {j = 1,2) is the matrix, the rows of which are the btJ component gradients calculated at the point y. Then linearization (22.16) at the point y0 yields:
[QiJiAvo)Ji2(yo)](y -- 2/Q) yo)>> Ml/o) bl2(vo)- QiKiVo) -QiKiyo)-■ [QiJitiVo) ~ JiM](y
(22.20)
Consider the coordinating problem which has convex objective function (22.18) and linear constraints (22.20): p
min f(y) = ] £ ( « £ )bh (y),
(22.21)
[QiJiiiVo) ~ Jt JiM](y > bl2i2{y (y0) - Qibi [QiJiAvo) Qib tfo), 2(yo)]{y - yo) > n(y0),
(22.22)
I = l,p.
Let y1 be the solution of (22.21), (22.22). Since y0 is feasible for it, then 1 fiy^^Hyo). /(if ) «£/(*).
(22.23)
Decomposition
Based on Separation
of Variables
147
It is possible that equality will occur in (22.23). In this case the current solution can be tested for optimally. T h e o r e m 22.2: (Optimality criterion). Let x°,i = T^ be the solution of (22.9), (22.10) for y = y0. Then nhe eecessary and sufficient conditions for ({x°}, 3/o) to be the solution of the original problem (22.3), (22.4) lies in the fact that yQ is the solution of the coordinating problem and the following conditions are fulfilled *• < ==> 0, o,
iZ==Tj), I,P,
(22.24)
where u}2 is a vector of the dual variables corresponding to the ith set of constraints (22.22) in the optimum of the coordinating problem. Proof. Sufficiency: When carrying out propositions (22.1) - (22.4), the solution is optimal for the original problem if and only if the Kuhn-Tucker conditions are fulfilled. We write the problem in the divided form p V
(22.25)
m i n2_\ ^ Ccii xii t >, min i=i
(22.26)
xi>b> AllnXi li(y), b^iy), A2ix iXl1>b >hi22(y), (y),
(22.27)
i = T~p. l,p.
Let un, «i2 be the dual variables corresponding to constraint (22.26), (22.27). Write the Lagrange function p
£ = ( C llxxl t + (btlh(y) (y) C = ^^2(c +uunh(b
- AiltlxXi) + ui2{bZ2(y) - A,A2xi2%x,)) )). i)
(22.28)
i=i
Write the Kuhn-Tucker conditions at the point ({x°},yQ) A'nUn Ai^ii
(i = l,p):
+ A' ui22 = , + Ai2l2Ui — ClCi,
Ji, (lto)«*i + -4(3/oK 2a =°> = 0, 4,(yoK +4„(ito)«i «*i(*f*(jta) < ( M * o ) - AAi^i) nx°)
= 0, 0, == «u'ii2, (b (M » 0»)) --AA l2(y i2x°) l2x°t) =
Ai.xl h(yo), A„i! > > bMsto)>
A l2x° > i2(y0), A^a;? > bM3/o):
(22.29) (22.30) (22.30) (22.31) (22.31) (22.32)
w ul212 > >0,0 ,
(22.33)
tti, >>00..
(22.34)
148
Systems
Optimization
Methodology
The optimality conditions for linear subproblems are as follows u^= (=A{A" < r i 1 )')Cl' c i >> 00, , M3/ 0 ), Antlx° = 6i,(yo),
(22.35) (22.36)
bi2(y0). ^A2l2x°i * ? > Ms/o)
The Kuhn-Tucker conditions for the coordinating problem at the point y = 2/0 are obtained from the Lagrangian p
£ = £ { « ) M y ) + ( « i , ) M * > ) - Qi^(3/o) £ = £ { « )*•■« (») + ("*.)[**. (2/0) - QiK (3/0) 1=
1
(22.37) (22 37)
--(Q ( Q i-/, ^ i ((3/o)-^ » o ) - ^ 2 ((j/o))(y-j/o)]} »o))(y-»o)]} 1 I 2 and, for i = T^p, take the form " QHi <] ] + 4^((W ^J'h o(»o)[< K - Q 2 >oo) )<< ,.
(22.38) (22.38)
« ))((*M i M » o ) ) == 00 ,, * (y»o>)) - -Q QiM»o))
(22.39)
<«J>a 0> 0- .
(22.40)
Comparing (22.29) - (22.34) with (22.35), (22.36), (22.38) - (22.40) we find that only conditions (22.29), (22.30), (22.34) may not be satisfied. Comparing (22.38) and (22.30) involves our assumption of Uil
= u? u*% = «?, u% - Q Qiu\ , < 2,,
(22.41)
therefore (22.30) is satisfied, multiplying (22.41), by Ah obtain
4X+4, «<<=«. A ' n u*n + A ' h Qiu]2 = A ' n u°tl , since, due to (22.19), Q[ = (A^YA^, (22.29). Thus, if <>0, < , >0,
u°n = (A^)'ci, »i=T3, = l,p,
(22.42) (22.42) changes to (22.43)
the Kuhn-Tucker conditions are satisfied in ({*?}, y0), here «£7 uj2 are the Lagrange optimal multipliers, i.e. the solution is optimal. Necessity is proved in the same way. Consequence: The sufficient condition for ({z?}, y0) to be optimal resides in the fact that y0 solves the coordinating problem, and none of the linear constraints is fulfilled as equality.
Decomposition
Based on Separation of Variables
149
Proof: If y0 is the interior minimum if, due to (22.39), u\2 = 0, due to (22.41), u*k = u°it. Since < > 0, then ({xj}, j/ 0 ) is optimal. If the criterion of optimally is violated, then either f(yl) < /(«„) or f(yl) = f(y0) and some < | 0. Consider first the caseftf) 4) since the coordinating problem has been formed under the assumption of Anx, = bn(v) and linearization of constraints (22.16), then the objective function f(y) of the coordinating problem is equal to ip(y) only for such y that Ah is an optimal basis of a linear subproblem (22.9), (22.10). Actually we want to reduce ip rather than / . Further, due to linearization y1 may not satisfy (22.16), and, thus, it may not be in S. The following theorem deals with these difficulties. It shows that either y1 € 5 and satisfies iply1) < tpiyo) or, if the nonzero step 0m can be accomplished along the line from y0 to y1 then it is possible to construct a new vector vn + 0 (v1 — vn) that exhibits the same property. Theorem 22.3: (Substantiation of linearization). Assume that an optimal solution of the coordinating problem satisfies the condition f(y11)
(22.44)
If y1 satisfies the conditions 1 l 1 hhiiy =lbQib^y ) {y1)>0, - b^iy1) > 0, l{y ))=Q n(y )-bl2
ii=Xp, = l,p,
(22.45)
then V'(?/ 11)(l/1)(2/o) = ^^ oo )) ..
(22.46)
If y1 does not satisfy (22.45) we will introduce 0 m as the largest step from y0 to y\ where the feasibility is retained: 1 max{e\hi(w o + = Up, 0 m = max{0|M3/ + ©(l/ ©(»* -~ Vo)) > > 0, i = l,p, 0 < 0 < 1} ..
(22.47)
Although if 0 m > 0, then ip(yo + em(y1-yo))(yo)^(yo Qm(y1-yo))<^(yo).
(22.48)
Proof: Inequalities (22.45) constitute, due to (22.12) - (22.16) the nec essary and sufficient conditions for the basis Ah being feasible for the ith
Systems Optimization Methodology
150
linear subproblem with y = y1. Since A^ may not be an optimal basis when y = y1, then M1) ("*).
(22.49)
Since An is the optimal basis with y - yo, then f(Vo) = *{Va) ■
(22.50) (22-50)
Combining (22.49), (22.50), (22.44), we obtain (22.46). To prove (22.48), we set
y = 2/0 + @m(j/1 -yo)Since / is convex, then /(y)(yo) + em(/(y1)-/(yo))-
The basic matrices An may not be optimal in (22.9), (22.10) when y = y so that V(y) < /(»)• Using (22.50) obtain
V'(y)
/(y 1 )(yo),
em >o,
so that ip(y) < ip(y0) and this is exactly (22.48). Consider the situation when fiy1) ) = /(yo) and a particular w^ ^ 0. Here the result is similar to that described in the preceding paragraph. T h e o r e m 22.4: (Alternative optimal basis). If /(y 1 ) = f(y°) ) and for a particular i, u*t there are some negative components, the optimal solution
Decomposition Baaed on Separation of Variables 151 exists for the ith linear problem (y — yo) such that at least one row of the original basis AZl corresponding to the negative component u* is absent. §22.2. Formulate the algorithm of the above method for solving problem (22.3) - (22.4). Step 1: Choose the vector y0 € S. Step 2: Solve subproblems (22.9) - (22.10) with y = y0, obtain the optimal solution x°, the basic matrices Aix and the dual variables u° Step 3: Construct the coordinating problem (22.21) - (22.22), solve it, ob tain the optimal solution y^.
Step 4: If f{yl) = f(y0) and all u*h > 0 (22.24), the solution ({x°}>2/o) is optimal. Go to step 8.
Step 5: If f{yl) = f(y0) and u*h ^ 0 for i € /, then by theorem (22.4) there exists an alternative optimal basis for the ith linear subproblem in which at least one row AH corresponding to the negative compo nent u* is absent. Such a basis can be found by solving the linear problem (11.19), §11. We will carry out at least one alteration in the basis and go to step 3 with a new basic matrix and dual variables.
Step 6: If fiy1) < f{y0), all h^y1) > 0, then due to theorem 3, ipiy1) < ip(yo). Suppose that yo = y1 and go to step 2.
Step 7: If fiy1) < f{yo) some hiiy1) "£ 0, calculate
The following cases are possible here:
(1) 0m > 0. Due to theorem (22.3), tp(yQ + Qm{yl ~ yo)) < i>{Vo). Return to step 2, set y0 = y0 + 0 m ( y 1 - Z/o)-
(2) 0m = 0. For some i,j there exists hij(y0) = 0, ^(y1) < 0. It can be shown that the row j of the matrix Ai2 can be replaced by a row A^, which leads to a new optimal basis in the zth subproblem. Using the new basis go to step 3. Step 8: Terminate.
152
Systems Optimization Methodology
Prove the statement made in step 7, in case 2. From ft»i(l/o) = 0 it follows that
where A* , b3^ are the jth rows A,2, 6i2. By virtue of assumption 3 A\2 ^ 0,
vso that A\ A~x ^ 0, i.e. A\ can be replaced by a row Aix, which leads to A^ a nonsingular matrix. The relation
(22.51) must hold and, since A^ is nonsingular, x° is a unique solution (22.51). The dual problem objective function C{Xl does not change, so that A^ is a new optimal basis. §22.3. The optimal algorithm finite convergence is proved by the theorem. T h e o r e m 22.5: If optimal bases do not recur at the algorithm steps 5, 7, then an optimal solution is obtained for the finite number of iterations. Proof: At each iteration, where the optimality criterion is not satisfied, the
coordinating problem objective function f(y) decreases or the optimal basis
is found in one or more subproblems. Whenever f(y) decreases we have to turn to a new set of optimal bases {A^} differing from the employed in any preceding cycle because iteration of the set will bring about iteration of the coordinating problem and, therefore, that of the minimal value of / ,
as well. Consider the cycle at which f(y) remains the same. We have a finite number of optimal bases for every y value. If none of these bases recurs at steps 5 and 7 of the algorithm, then at each iteration the current set of optimal bases cannot coincide with the set existing at the preceding iteration. Since the complete number of permissible sets {A n } is finite, then an optimal set can be obtained for the finite number of iterations. §23. B E N D E R S METHOD FOR A SPECIAL MATHEMATICAL P R O G R A M M I N G PROBLEM §23.1. The Benders method based on variable division deals with the ma thematical programming problem of the form
Decomposition
Based on Separation of Variables
min{cr+ /(»)} ,
153
(23.1)
Ax + g(y)>b, (23.2)
(23.2)
x > 0, yeS, (23.3)
(23.3)
where c,x are n-dimensional vectors, y is a p-dimensional vector, f(y) a scalar function, b an m-dimensional vector, g(y) an m-dimensional vector
function, A is an (m x n)-dimensional matrix, S a particular subset of Ep, e.g. a set of integer vectors. Since problem (23.1) - (23.3) is linear by x with the fixed y, it is natural to attempt solving it via fixing y, solving the linear problem with respect to x, obtaining a "better" value of y, etc. Let (23.4) R={y\3x>0:Ax>b-g(y), y £ S} . (23.4) The vectors y £ R will be called feasible. The set R can be specified explicitly, for which we will employ the Farkash lemma. The Farkash lemma. There exists the vector x > 0 satisfying the condi tions Bx = a if and only if a'u > 0 for all u satisfying B'u > 0. Fix y and apply the lemma to the linear problem.
Ax- s = b-g(y),
x>0, s > 0.
Hence it follows that y is feasible if and only if the following condition is fulfilled (b-g(y))'u<0 (23.5) (23.5) for all u satisfying Since the cone
A'u<0, w>0. C = {u\A'n < 0, u > 0}
(23.5')
is polyhedral, then it is determined by a finite number of generatrices,
i.e. the vectors u\ (i = l,nr) exist such that any element u 6 C can be presented in the form (23.6)
154
Systems Optimization
Methodology
Substituting (23.6) in (23.5), obtain
f>(6-0(y)K
which holds for all Ai > 0 if and only if T (6 - g(3/))'< (b-g(y))'u l<0, < 0,
= T^r l , nT.r . iz =
(23.7)
Thus, the set R in (23.4) can be represented in the form {y\(b - 9(y)Y< R = {y\(P g(y)Y<
y € yes}. S}.
If R is empty, then the original problem (23.1) - (23.3) has no (feasible) solution. In case R is nonempty, (23.1) - (23.3) will be rewritten as mm{f(y) m'm{c'x\Ax g(y), xx >> 0}} 0}} min{/(j/) + + min{c'a:| Ar >b > b— - g(y), y€R
(23.8)
For the fixed y, the interior minimization of (23.8) is the LP problem. We will write this problem and its dual problem. The direct problem min ex, Ax>b-g{y), The dual problem
(23.9)
x>0.
max(6 g(y))u, m a x ( 6-- g(y))u, A'u < c, u > A'u
(2^1a) (23.10)
According to the duality theorem the following equality of the optimal functionals of problems (23.9), (23.10) exists: mm{cx\^x
> b-g(y), x > 0}=max{(6-g(y))u,
A'u < c, u > 0} . (23.11)
Substituting (23.11) in (23.8), we come to a new form of the original problem min{f(y) mia{f(y) + max{(6 max{(& - g(y))'u\A'u g{y))'u\A'u < c, u > 0}. 0}. (23.12) y€R y€R
Consider the constraint set of the dual problem (23.10) P = {u\A'u < c, u > 0}
(23.13)
Decomposition
Based on Separation of Variables
155
The set does not depend on y, and the magnitudes < in (23.5') are the P set extreme rays. If P is empty, then the criterion value for prob lem (23.9) and, therefore, that for problem (23.1) - (23.3) are not bounded below. If P is nonempty, then the interior maximum in (23.12) is achieved at one of the extreme points of P or approaches oo when moving along the extreme ray of P. In the latter case, direct problem (23.9) is unaccept able, which contradicts the initial assumptions so that we confine ourselves to considering only the extreme points of P And such points constitute a finite number Denote them b y / i = T I T Problem (23 12) is now rewritten as (23.14) min{/(i/)+ max (b - g(y)) V) . min{/(i/) + max(fe g{y))'u\} y£ V£RR
i=l,nTp i=\,n
And the problem is equivalent to the following min z, (b- g{y))'u g(y)Yupti,, z > f(y) + {b-
= TjT i= l,np p, ,
(23.15)
yER. Using the definition of the set R, (23.15) is written as min z, p (b- g(y))'u z > f(y) + {bg{y))'uvt,,
(b{b- g(y))'ui, g(y))'ui,
i = T~np~, l,np ,
i = i~l,n l,nrr,,
(23.16) (23.16)
yes Thus, the relation between the original problem and problem (23.16) is clear. These results are summarized as the theorem. T h e o r e m 23.1: (1) Problem (23.16) has a feasible solution *> (23.1) - (23.3); (2) if (z°,y°) is a solution to (23.16), then (x°,y°) is the solution of original problem (23.1) - (23.3), where x° is a solution to (23.9) with y = y°; (3) if (x°,y°) is a solution to (23.1) - (23.3) and z° = c'x° + f{y°), then (z°,y°) is a solution to (23.16). Unfortunately, the process of solving problem (23.16) is made difficult due to an enormous number of constraints. Therefore it is natural to
156
Systems Optimization
Methodology
apply the relaxation procedure described in §9. We make the following assumptions: (1) 5 is closed and bounded; (2) f(y) and the components g(y) are continuous on S. L e m m a 23.1: If assumptions (1), (2) are fulfilled and (23.16) is feasible, then z has no finite lower bound & P is an empty set. Proof: See Ref. 19. Consider problem (23.16) on a constraint set min 2, z, {b-g(y))u i € i€h, h , zz>f(y) > f(y) + (bg(y))'ul l, {b-g{y))'u\
(23.17)
where h C { 1 , . . . , n p } ; h C { 1 , . . . ,nT}. When satisfying the assumptions (1), (2), if (23.17) is feasible, then it has the finite optimal solution (z°,y°). The solution is optimal for (23.16) <£■: 0 (b-9(y ))'ur
( 6 - 5<(/t{/ /°) ) ' < < 0 ,
= T^T l , nPp, , iz =
i= = TT^T ^ 7 r..
(23.18) (23.19)
In order to get rid of a large excess, we will introduce the auxiliary linear problem max(6-g(y°))uu (23.20) (23.20) uu €e P. P. The index of the extreme point, at which solution (23.20) is achieved, co incides with the index of the most unsatisfied constraint (23.18). Problem (23.20) coincides with dual problem (23.10) with y = y° Since 2° is as sumed to be finite, then, due to lemma 23.1, P is nonempty. Thus, (23.20) has a finite optimal solution or an unbounded solution. In case it is not bounded, the objective function tends to +oo along the half-line upt + A<, A > 0. In this case (b - 3<7(2/°)K (2/°))'< > 0
Decomposition Based on Separation of Variables 157 for a certain i. Hence it follows that one of constraints (23.19) is violated. Thus, (23.18), (23.19) is satisfied by O :
max{(6 - g(y°))'u\A'u
then it is optimal for (23.16) <=> :
max{(fr - g(y°))'u\A'u
max{(6 - g(y0))'u\A'u < c, u > 0} > z° - f(y°). (23.23) (23.23) If the maximum in (23.23) is achieved at the extreme point u° then
(b-g(y0)Yu0>z°-f(y0). (23.24)
(23.24)
(23.24) establishes the new constraints that should be introduced into a new problem (17). If linear subproblem (23.20) is unconstrained, then the simplex method
leads to the extreme point P - u° and the extreme ray P denoted by v° so that (b - g\y0))' approaches +oo along the half-line u = u° + Xv°,
A>0
(23.25)
and, in order it might occur, v° has to satisfy the conditions (b-g(y0))'v0>0, (23.26)
(23.26)
which prescribes a new constraint in form (23.19) to be added to (23.17) for its new formation. §23.2. Consider in greater detail the solution of problem (23.17).
158 Systems Optimization Methodology Although the above procedure can be formally applied to a very wide class of problems, in practice it is efficient when (23.17) can be efficiently solved. Consider the case when it is possible.
(1) 5 is a set of vectors in Ep with the non-negative integer components g{y) — By, f(y) = d'y. In this instance, (23.17) is an integer linear problem. And since sequential problems (23.17) differ only in adding one or two constraints, then, for their solution, the cut-off methods of the Gomory method type are appropriate. (2) S is determined by a set of linear and nonlinear inequalities
where gt are nonlinear continuous differentiable functions. The func
tions f(y), g(y) can be nonlinear, and also continuous and differen tiable. Then (23.17) has a linear objective function but nonlinear constraints. Whatever the applications may be, the Benders procedure displays the ability to maintain any special structure of the matrix A. For instance, if the matrix A is of a transportation type, then (23.19) also is a transportation problem in which right-hand parts change at each step. And since at each iteration of (23.19) only right-hand sides change, the optimal dual solution at iteration i remains feasible for a new dual problem at iteration i + 1, as well. Thus, the dual simplex method is supposed to be the most suitable for its solution. §23.3. Formulate the Benders method algorithm for solving problem (23.1) - (23.3). Step 1: We start the procedure with permutation of problem (23.17) in which there are several solutions. Step 2: Solve problem (23.17). If it is unfeasible, then so is the original problem. Otherwise, we will obtain the finite optimal solution
(z°,y°) or data on solution unboundedness. If z° = -oo, then the
y° derivative element is taken as S. Go to step 3. Step 3: Solve dual problem (23.10). If it is unfeasible, then, due to lemma 1, the original problem has the unconstrained solution. If the dual problem is unbounded, then go to step 6.
Decomposition Based on Separation of Variables 159 Step 4- If the objective function optimal value in step 3 is equal to z° —
f(y°), then the solution of (z°, y°) constitutes the solution of
(23.16). If x° is the solution of (23.9), then (x°, y°) is the so lution of (23.1) - (23.3). Go to step 7. Step 5: If in step 3 the criterion of optimality is not satisfied and dual problem (23.10) has the finite optimal solution u°, then
z0
(23.27)
Add this constraint to (23.17) and return to step 2. Step 6: If (23.10) has the unconstrained solution, then the simplex me thod enables one to find the ray v° and the point u° such that the objective function (10) tends to +oo along the ray: u = u° + \v°, A>0. (23.28) The inequality
(23.28)
(b-g(y°)yv0>0
is satisfied so that y° does not obey the constraint
(b-9(y))'v°<0. Add this constraint to (23.17). If
Z°
160 Systems Optimization Methodology been obtained or with information on unfeasibility or unboundedness of this problem. Proof: Problem (23.16) has only a finite number of constraints. If the optimality criterion is not satisfied, then one or more new constraints are added to (23.17). Therefore, for the finite number of iterations the optima lity criterion is satisfied or the full set of constraints is employed. (23.16) is unfeasible if and only if (23.17) is unfeasible at a particular step. (23.16) is unbounded <$> dual problem (23.10) is unfeasible. The algorithm features the possibility of calculating the upper and lower bounds of the optimal objective function value. The upper estimate is gen erated by the sequence of feasible solutions (23.1) - (23.3) so that the better of these estimates can be taken as a solution if the procedure terminates before an optimum is achieved. Assume that (z\j/*) are the solutions of (23.17) at the step i and linear
problem (23.9) has a feasible solution with y = yl If xl is the solution of (23.9), then (a;*, y*) is feasible for (23.1) - (23.3) and the following inequality exists
cxl + f(yl) > mm{z\(z, y) £ G) ,
(23.29)
where G = {(z,y)\(z,y) obeys the constraints of problem (23.16)} obtaining the lower bound follows from the fact that constraints are added to (23.17) so that if 21 is the objective function optimal value for (23.17) at step i, then
zl < zi+1 < min{z\(z,y) € G}
(23.30)
Conditions (23.29), (23.30) at the step i give the estimates (23.31) With the iterative procedure converging, the right- and left-hand sides of (23.31) converge to the value confined between them. And the optimality criterion represents the condition for the upper and lower bounds being equal. §23.5. We will study the relation between the decomposition methods of Benders and Dantzig-Wulf.
Decomposition
Based on Separation
of Variables
161
Consider instead of original problem (23.1) - (23.3) the LP problem:
min(ca; + dy),
Ax + By>b, x > 0,
(23.32)
y >0
and the problem dual with respect to it: max bu, B'u
(23.33)
Assume that the set P= {u\A'u<
c, u > 0 }
is bounded. Then problem (23.16) for (23.32) takes the form mm z, z + {By - b)'uj > dy, Vj ,
(23.34) (23.34)
y>o, where Uj is the j t h extreme point of P Problem (23.17) for (23.32) has the same form, though only a certain subset of constraints with j numbers
are taken into consideration. The optimality criterion of the point (z°,y°) constitutes fulfillment of the inequality
max{(6 - By°)u\A'u < c, u > 0} = z° - dy° ,
(23.35)
where (z°,y°) is the solution of problem (23.34) for a certain set of con straints with j numbers. To test the criterion, we need to solve the linear problem:
max(6 - By°)u,
A'u
(23.36)
u> 0.
We apply now the Dantzig-Wulf decomposition principle to problem (23.33). Write any u € P as a convex combination of the extreme points ofP:
(23.37)
166
Systems Optimization
Methodologo
substitute the result in the remaining constraints of (23.33) and its objective function, which leads to the coordinating problem: max 2~](buj)Xj 2~](buj)\j ,
Y^iB'u^Xj
(23.38) (23.38)
j
3
J > = 1, 3 J
Xj > 0. The problem dual with respect to (23.38) has the form: min(dy + v), (23.39)
v + (ByYui > bu, , y>0~ or, setting
(23.40)
z = dy + v we get
min min z, z, zz + + (BybYu, iBy-bYuj^dy, > dy, 22 > > 00 ..
Vj \fj
(23.41) (23.41)
Problem (23.41) coincides with problem (23.34) in the Benders method. Let (y°,v°) be the dual variables corresponding to the optimal solution of the reduced coordinating problem. Consider the local subproblem of the decomposition principle 0 max(b-By°Yu, max(b-By Yu, k0. u>0.
(23.42)
It is clear that (23.42) coincides with subproblem (23.36) in the Benders algorithm. The optimality criterion of the coordinating problem is as follows max{(b6-y m a x { ( 6 60)u\A'u|A'«
u u uu0}0}< 0} 0}<
(23.43)
Decomposition
Based on Separation of Variables
163
It has been shown above that the inequality is actually fulfilled as equality. (23.40) implies
z°=dy°+v0
and (23.43) transforms to the form
max{(6 - By°)u\A'u < c, u > 0} = z° - dy° , which coincides with the optimality criterion in the Benders method. Thus the methods of Benders and Dantzig-Wulf are dual to one another. The Benders algorithm is applied, however, to a wider class of problems, e.g. the partially integer problems of mathematical programming. §23.6. And as conclusion of the paragraph we will consider one search algo rithm of approximate solutions for the following partially integer problem 18 : max{/i(x) + / 2 ( y ) } ,
< b,
(23.44) (23.44)
y£Y,
where x is the n vector, y is the integer m vector, b the / vector, Y the bounded set, f2(y), 1^2(2/) the arbitrary functions, and fi(x) the twice con tinuous difFerentiable concave function. Assume that the feasible set of problem (23.44) is bounded and nonempty. The maximum of the Euclidean norm of the vector x on the set is denoted by M. Formulate the search algorithm of the approximate solution of prob lem (23.1) based on the Benders decomposition concept. In the algorithm
description we will meet two sets of vectors {Pi},i = 1,2,..., Tk, {A.,},
j = 1,2,. -.,qk, three sets of scalars {pi},i = 1,2,..., Tk, {fij}, {A.,}, j = 1,2,...,qk, where k is the number of iterations. Step 1: Choose the vector y\ £ Y and the number di being larger than the objective function optimum value of problem (23.1). Set k = 1, r0 — 0, qo = 0. Introduce the accuracy parameters of the scheme £i,i = T75 the meaning of which will be clear from the subsequent steps of the algorithm.
164
Systems Optimization
Methodology
Step 2: Solve the problem mmip(x,yk,dk), miaip(x,y x X
i>(x,yk,dk)
ff22(yk)}) (yk)})2
= {max{Q,dk - fi(x) h(x) i
+ ^ ( m a x { 0 , (nix)
+
!=1
= =
2 \dk\dk-h{x)-f -h{x)-h{y )]l 2{yk k)\ +
+ \\{v \\(Vi(x) 1(x) +
2 ip22(ykk)-b) )-b)+\\ +\\ E.
In this case, it is supposed that the desired minimum is found with sufficient accuracy, when we obtain the point xk in which even one of the following three situations is realized. Case 23.1: ^{xk,yk,dk) < et. The pair xk,yk is taken as the approximate solution of the original problem. Go to step 5. Case 23.2: il>(xk,yk,dk) > elt l/2\(^'x(xk,yk,dk),xk)\ < e2^{xk,yk,dk), \l2U'x{xk,yk,dk)\\B < e2[dk - h{x) ~ hiVk)]In this case we set rk = rfc_i + l,qk= g fc -i, the vector pT. = [
+ h(xk)
- (f[(xfi( b k),X k) Xk)-(ft{x k),X-k)-b
into the set {Pl}. Go to step 3. Case 23.3: i>(xk,yk,dk)>eu \\[
v[(xk)xk). ti(xk)xk).
In this case we set qk = qk^ + l,rk = rk-lt the vector \qk = [ipx(x) + *Pi(Vk) ~ &]+ is introduced into the set {A;}, the value Aqk = (A w , b
Decomposition
Based on Separation
of Variables
165
Step 3: Solve the problem of the form My)} max{ min_fa min fa - (pi,
(23.45) (23.45)
In the case of unsuccessful choice of the parameters e 4 , e 5 , the feasible set of problem (23.45) can prove to be empty and calculations terminate, though the approximate solution of problem (23.44) has not been found. Step 4'- Let yk+i be a solution to problem (23.45). Define the value dk+i = min{dfc,/2(i/jfc+1) + min fa dk+i = min{dfc,/2(i/jfc+1) + 1 =min_fa 1,Tfc -
(pi,
i=l,Tk
and form the function i{>(x,yk+i, <4+i). Go to step 2. Step 5: Terminate. The algorithm convergence to the approximate solution of problem (23.44) for the finite number of steps is shown by the following theory.18 Theorem 23.4: Proceeding from any initial approximation and with 0 < £4 < £5 < 1/M, 0 < £3 < 1, £1 > 0, £2 > 0 the algorithm described will find the approximate solution of problem (23.44) for the finite number of steps. Here directing £i,£2 towards zero we can provide any degree of closeness to the true solution. If we assume that the problem mintl>(x,yk,dk) is accurately solved, i.e. £1, £2 = 0, then the algorithm yields an exact solution to problem (23.44) either in the limit or for the finite number of steps (here the finite convergence occurs in the linear case).
Chapter 6 DECOMPOSITION BASED ON OPTIMIZATION TECHNIQUE
The present chapter deals with application of numerous well-known methods for finding the dynamic programming extrema, the conditional gradient, the componentwise descent, etc., to the problems with a spe cific structure, which enables one to carry out, to a certain extent, their decomposition. §24. A P P L I C A T I O N O F T H E C O M P O N E N T W I S E D E S C E N T M E T H O D F O R SOLVING T H E PROBLEMS OF MATHEMATICAL P R O G R A M M I N G
AND OPTIMAL MANAGEMENT §24.1. We set forth first the componentwise descent method. Consider the problem min/(x) = f(xi,...,xn), (24.1)
(24.1) (24.2)
where M* are the bounded line segments. The idea of the method resides in the following: if the approximation
xk = (x*,. o o ,x^) is formed, then the following approximation is also deter mined by solving the minimization problems n (i = l,n) with one variable 166
Decomposition Based on Optimization Technique k+1
min m i n f(x f[x1
k k k ,...,x , . .^,Xi,x . , !B;_i ,Xi,Xi+ i, +1,...,x n),
...
,xn),
167
(24.3)
& = 0 , 1 , . . . ; x ° , . . , , x°n are initial approximations. The method convergence has been proved (see, e.g. Ref. 4) under as sumption that f{x, is the convex differentiable function in the region in volving the closed, boundeedonvex set Mx x M\ x • o • o Mxn. In .his casc there exists the equality lim /f{x ( x kf )c ) == min f(x). f(x). fc-»oo k—►oo
(24.4)
ix£M £ M xz
§24.2. With a particular form of functions (1), the componentwise de scent method generates the hierarchic optimization scheme. In Ref. 35 we consider the problem m
min £ / < ( * * , « ) ,
(24.5) (24.5)
t=i i=\ l xttEM €M i = l,m, u€v. uev. x, x, i=Tjn,
(24.6)
By virtue of the specific structure of the objective function in (24.5), mini mization by xt (i = T~^) with the use of the method (24.3) can be carried out independently and simultaneously. Thus we obtain the following scheme for solving problem (24.5) - (24.6) at two levels: (1) Select the initial approximation u° Gv. (2) If the approximation xk = ( i f , . . . , « * » ) , uk is constructedd,hen nk+1 is constructed according to the rules: (a) solve the problems min fi(xi,uk), i~\,m min/<(*«,«*), i = T^i
(24.7) (24.7)
the solutions of which are taken as xk+11 (b) at the second level, we solve the problem 1 minV/^x^ ,^) minV/^+V) i=\ i=l
the solution of which gives ukk1.
(24.8)
168
Systems
Optimization
Methodology
The present approach can be generalized for constructing the optimiza tion schemes at several levels. For instance, we need to solve the problem m m,
(24.9)
m i n ]2J P ] r2J / i j (&(*«>*«»*) :Eij,:Ei,u), ■ min i=l ; = 1
xtj e€ Mi\ ary M*J, x,eM' x, £ Mxj,,
u eu£v. «.
(24.10)
In this case a new approach xk+1 = (s^f1,... , i ^ , i f + 1 , . . . is constructed as:
,xk^'1,uk+1)
(1) the first level subsystems solve the problems min
xij€M'x3
flj(xij,xk,uk),
l , m , j = 1,TH» i = l,m, l,Wi
the solutions of which are taken as a£ + 1 ; (2) the second level subsystems solve the problems m, 1 k min ^},fijixjf . / ^ ( a s * / 1 , *,%i,u. * , « *),) ,
i-l,m
from which we obtain xk+1; (3) the third level solves the problem 771 m
TTli TTli
i=l j = l
The process is iterated with k = k + 1. Evidently, the solutions of the problem m m i n
mi,
^ ^ - - -
m; 1 ...t„_2 mi,...t„_2 ^2
f'i-irt-l(Xh...in.1,Xlli2,Xil,u),
1 iEij.-.i, iEij.-.i, € M A J *£ 1 ' '" -- '^* '%, . ... ., I^I S£ v) , jZ,=! =l ,l n, n- - 1l
with the aid of (24.3) we obtain the scheme of optimization at n levels. In conclusion it should be noted that problem (24.5) - (24.6) has an economic application.
Decomposition
Based
on Optimization
Technique
169
Assume that we have m objects (subsystems) utilizing various (their own) resources xt E Mx(i = T~m) and the common (central) resource u = m + • • • + um £ v. The profit of the fth subsystem is /*(£,-,«»), the profit of the whole system is £ £(*<,«*). The solution of the problem at two levels: max. max, fi(x / t (it*Ui) i , t i f ) - + xa £- + 1 ;
(1)
771 m
1 m m aa xV^ // i. (fa^; J' +. 1u, «, )i )-- ,« ^* + 1. .
(2)
Similarly, it is possible to formulate the problems leading to the schemes of optimization at many levels. §24.3. Consider application of the componentwise descent method for op timization at two levels of a discrete dynamic system involving m centre controlled independent subsystems: m
N —1
m l
+ ^2fiN{xlN)
minJ2^2 m i n ^ ] P fik(xl,u (xi,ui,x ,ukk)) k,xkk,u i=l
lc=0 k=0
+ fN(xN),fN(xN),
(24.ii;
i=l »=1
x
4k+i +i = + DDlu{ G\ukk,, = C4\xl4 + lul ++ G\u xk+ii=c=c xk ku+D kxkk+D k, kuk,
(24.12J
xi x\
x0(k(k = 0,N-l), 0,N -1),
u\. 6 vx;uk 6 v,v,
(24.13)
where x\ are the phase variables of the ith subsystem, xk are the phase variables of the centre, ul the ith subsystem "own" controls, uk "central" controls. Denote by
i? G {t ,t ){D)u)m+ + G)K)'), xi=
(24.14) (24. 14)
X = 4>(t ,t00)x xkk=4>(t +0 + k,t0k)x
(24.15)
z{ = ^(*fck ,*o)4 0 0+
z
k j+l
tn
22 ^ t0
4>(tk,tJ+1
Uj
170
Systems Optimization Methodology
Substituting (24.14), (24.15) in (24.11), we obtain the minimization prob lem of form (24.5) -(24.6):
ii
m N-l m N-l
( /
^ ^ / , * U,(tk,t0)x1Q+ i= l
fc=0
i 00*(*fc.*i+l)(^«i (*fc,«i+i)(^«i
Y
V
ulk,<j>(tk,t0)xo+
Y
i i)'
4>{tk,tj+i)DjUj,uk\
t0
f}
+ YfmU'(tN,to)xl+ i=l
+ G U u
to
Y, Y
\\
t0
IN I
)
^Nitj+^DjUj J> .
Here, in (24.5) - (24.6), xj correspond to u{, and u are u^ equations. §25. CONDITIONAL G R A D I E N T METHOD A N D DECOMPOSITION OF PROBLEMS OF MATHEMATICAL P R O G R A M M I N G A N D OPTIMAL CONTROL §25.1. Consider the convex programming problem min/(x),
xSX,
where f{x) is the convex functional, X is the closed, constrained con vex set. We set forth briefly, for this problem, the conditional gradient method in which the following iteration process is formed: (1) select the initial approximation x° € X; (2) if the approximation xk is found, then the following approximation xk+l is constructed according to the rules: (a) solve the minimization problem rain{gmdf(xk),x); i ) ; min(grad/(a;*), X6-A
(25.1)
Decomposition
Based on Optimization
Technique
171
(b) xk+1 = xk + ak{xk - xk), 0 < afc < 1, where xk is the solution of (25.1) (fc = 0 , 1 , . . . ) . There exist various possibilities for select ing the step ctk according to which different versions of the above method are obtained. The method convergence has been proved under the assumption that f(x) is differentiable, and the gradient grad/(x) satisfies the Lifshitz condition (see, e.g. Ref. 4). Then it is proved that with the appropriate choice a t k /(**) (1) lira lim f(x ) = = /* f*== mi mif(x); f(x); fc—►oo k—*oo
xx ££ XX
(2) f(x f{xkk) - /f** < < §, where c -= const > 0; (3) there exists the sequence xk converging to the minimum point x". §25.2. Consider application of the conditional gradient method for decom position of large nonlinear programming problems (26). Assume that we have the problem min f(xu...,xf(xi,...,x n), n),
(25.2)
where f(x) is convex, Xi (i = l,n) are the closed bounded convex sets. The very form of problem (25.2) suggests that in the system under consideration one can isolate n subsystems each of which is in agreement with the vector X{. We will construct the iteration process for solving (25.2) based on minimizing some objective functions for each system separately (by every vector Xi). Denote grad/(x) {Tx(x),...,T grad/(a;) = {Tx{x),..., Tn{x)). (x)). Since in the present case n TX
k
(gra,df(x ),x)
Y,(K(xk)^r),
= i=i
then problem (25.1) falls into n independent minimization problems min (Ti(x ),xll),
i=
l,n.
x,6A\ x,6X,
The given circumstance enables one to organize the following solution pro cedure for problem (25.2) at two levels: (1) selects 0 = ( « ? , . . . , * £ ) , x° k
(2) if x
is found, then
eX,;
172
Systems
Optimization
Methodologl
(a) the second level calculates Firxk), transfers it to the tth subsystem (i= M ) ; (b) the ith subsystem solves the minimization problem k k min mxD.{F(Ti{x i{x ),x),Xi), i),
(25.3)
x,€Xt
the solution of which is 5*; (c) by f*, calculate a new approximation from the formula k+l k k xk+1 =x=x +ak(xk-xk),-xh),
(i = T7n), l , n ) , fclb = 00, ,1l ,, ......
(25.4)
It has been noted above that different possibilities exist for selecting the step ak (see, e.g. Ref. 4). The two-level procedure described provides an example of decentralized planning. Every subsystem i — M reports its plan xk to a central unit at the fcth moment of time. Using the plans obtained as the base, the central unit formulates sandeports to subsystems T,(xk)x i.e. ge.eral directions of changing the plans formed by them. Thereupon every subsystem i derives its new plan xkk1, reports it to the central unit ana so on. §25.3. We apply the conditional gradient method to decomposition of large problems of optimal control when the equations of moving with respect to phase coordinates lead to the block-triangular form36:
m i nf(u)= min /(«)=/
Jr(x{t),u(t),t)dt {x(t),u{t),t)dt
+ (j)(x(t)), <j>(x(t)),
(25.5)
Jt0
Ui(t) € in m
(25.6)
the phase variables x(t) = (x^t),....*„(«))),he eontrols u(t) = un(t)) are related dy a system of linear equations nn
nn
ii(t) = Au(t)x + Y^ Y, Aij(t)xj >M0zi + X)£ii(*K, ±i(t) =A ^2 Bij(t)uj , ii(t)xl t + j=t+i
xl(t00)) = x10 Xi(t to,,
iz = T~H l,n
(Ul(t),),.,
(25.7)
j=i
(25.8)
Decompositios
Based on Optimization
Technique
171
where Xiit)) ut(t) )re ehe eectors of ofe m dimension and the matricer AZJ, Bn have the (n, x n3) dimension. Assume that vx ara tht closeds bounded convex sets. The formula grad //' = = Tu Tu- - BTB $,T^,
(25.9)
where ip(t) is a solution of the adjoint system
(25.10) 3-1 i-l
in(t) = - 0 I f ( x ( T ) ) , S t) = i>i(t)=-Kwn), BTr ({t) = (J5$(i)), {Bjl{t)),
(25.11)
is valid. To implement the conditional gradient method, we have to calculate the grad / with the fixed u(t) = uk(t), and to do this requires the appropriate x(t,, ip(t). Since systems (25.7), (25.10) are block-triangular, then in the present case the linear inhomogeneous systems are successively integrated for finding x{(t), ip^t) Xz + ±i~ = AaXi + fi(t), r i>i = i>i = AliPi+ Alipi+g 3i(t), i{t)1
(i == n, ^xl(t W0) - S=i xoio (t n , .. .. .. ,, 1l )) ^, ^(T) iPt(T) =-4>' =-
(25.12)
where n
fl{t)=
n
^
^^(t) + ^^w,(t),
ii=«+i =«+i
y=i j=i
it -- li
It is known that *,-(*) = T T,(t)x + Ti(t) f T-\T)ft(T)dT, Xi(t) = t(t)xl010 +
(i = n , . . . , 1),
Jt ■/to 0
l l f.(t) =-[if = -\T?(t)]TT{T)4> Vl{x{T)) ^(t) (t)]rf (r^zcr))
[Tf (t)](tT1 1 yJ T?(T)g>(T)dT, (l = = 1, ... ,71) + [2f If(r) 9 i(r)dr, (i 1,...,n)
(25.13;
174
Systems Optimization
Methodology
where Tt (z = hn) are fundamental matrices for the system ±i = JA4li1Xi(t), iXi(t), i = l , n . From formulae (25.13) it is possible, given 7}(*), 7 f (*), T ^ * ) , P f W ] _ \ to calculate successively xn,... , X i , ^ i , ? . . , Vn with the fixed o(«). We set forth a general scheme of solving (25.5) - (25.8) at two levels: (1) select the initial approximation Q uu°(t) (t) = = «(u° u?(t)€t7*, zi = = Tl,n, ( *1(t),...,u°Jt)), ) , . . . , < ( * ) ) , U?(t)€t7*, ^,
(2) if the approximation nu is found, then uk+i is calculated according to the rules: (a) from formulae e25.13) calcalate xk(t),^k(ip uk(t)) in the following order the nth subsystem the (n - 1) subsystem
the 1st subsystem
x
n(t)
V*n(0
Ii
T T
zJi_i(*)
(with the fixef uk d
V'n-iW ^n-i(')
1
T
II xk(t)
T T 4>i{t) ipi(t)
(b) by these functions, calculate from formula (25.9) k g™df(u )
where et(uk)
= TUt f
=
(G1(uk),...,Gn(uk)),
(B-TP)1;
(c) the subsystems solve minimization problems (and their solutions exist since a linear functional achieves a minimum on the closed, bounded, convex set) min(Gl(Gi(u min (uk),ukl), ),Ui), u,£v,
(25.14)
Decomposition
Based on Optimization
where G% = (Gn,...,
Technique
Gint n,); .); uu,t = {un,...,
((G G tt((UU*), * ) ,UUl )l ) = / Jt0
.. i=1
175
«;„,);
k Y^G^it^u^dt; Y,G%J{u {t))u^{t)dt;
denote the solutions of subproblems (25.14) by uu k{t),
* =i=T/n; i.»;
i(t)>
(d) by the uk(t) derived, calculate a new approximation k uk+1+1(t) = «f(t) uk(t) + a fck(u (t) - «?(*)), uk(t)), i = i=T^. u* («) = (flf(0 1,« •
Subsequently, the process is iterated (k = 0 , 1 , . . . ) . §26. U T I L I Z A T I O N O F A P E N A L T Y C O N S T A N T I N D E C O M P O S I T I O N OF T H E MATHEMATICAL PROGRAMMING PROBLEM §26.1. Consider the block programming problem N
minima;) = ^ / t ( x t ) , 1=1
(26.1) v
N
^2sk(xk) ^29k(xk)
Rk,
k=l,N,
i=l
here xk = (xkl,...,xk^)
e Es" (k =
x = (xi,...,xNN)
€ Enn
TJt), l n« = ^XSl fScf cj ) ,'
f ) gk(xk) is an m-dimensional vector function, fc=i
6b = = (b ( bui...,b , . . .n,),6 m ) > R i Jk t C EE's* (lb (* = 1,JV), hN), assume that Rk (k = UN) are convex, closed and bounded, F(x) is the N
convex function differentiate on Rx x ■ ■ • x Rn,
£ 9k(xk) is the convex
vector function which is continuous on Rx x ■ ■ • x Rn.
176
Systems
Optimization
Methodology
Introduce the variables u = (U\,...,UN), V = (v%, -.. ,t?#)i where Uk = (ttfcl ,.-.,ukm),vk = (t/fcj,..., v Am ), (fc = I 7 F ) , (u, u) € E2mN and consider the problem equivalent to problem (26.1): minF(i), gk(xk) -uk
< 0, k = 1,N, l,N,
xk & Rk, k = 1, JV, x/t € i?*, fc = T j V , JV
||tl - o | | i ^ r = 0, ^vk
(26.2) (26.2)
|M|2
where i? satisfies the condition N N
ma
R>Y1
? IM^'OHl*'
fc=l
Denote
G w + 1 = |i «v ^£ l >v*k < 6b,, <
(r= , m ; fck = l , 2 V )\L, < fRl (r = l T~^; = TJf)
G = Gi x x • • • GN x Gx+i Select a strictly increasing sequence of positive numbers {Mi}, for which M, —^oo and reduce solving problem (26.2) to solving a sequence of the problems (2/), ( / = 1,2,...): min{F(x) M,||u-i>|| min{F{x) + Mt\\u - v\\2}2},, {x,u,v)€G.€ G. (*,«,w) Describe the iterative process of solving problem (26.2): (1) Fix v'-1 (with h = l,v° is chosen nrbitrarily) ana select ectl(u11) G, x • • • x GV on which mm{T(x) min{F(x)
l + Ml\\u-v M(-1f} ||«-«'-1||2}
€
(26.3)
is achieved. And to do this requires solving N convex programming prob lems of the form: min{/ fc (i fc ) + Mt\\uk - vl^l\\2} ,
(x£',»i') (xt',*i') e G*.
Decomposition
Baaed on Optimization
(2) Fix the solution of (26.3) (xll,u11) which a minimum is achieved
Technique
177
and define the point v11 € GN+1 at
r(xu)+Mi\\ull-vf.
(26.4)
And the problem is expanded into m subproblems: derive (vi\ , . . . , VrN ), at which m
N N
™Y^(ul'k k=i h=1
-vTh)2, (26.5)
N N
$>**
(k=TJ?), Y,vr
is achieved, here r = Vm. Problems (26.5) constitute strictly convex square programming problems. Fix v11 and go to point (1). In this way one forms the sequence {xil,ull,vu} involving a finite or infinite number of convergent subsequences. Isolate among these sequences the one converging to (x',u',vl). Let M21 be a set of all solutions (2/). Denote A ; = ||o' - v'\\2, x, ul, vl is a solution to (2/) and show that for all (x,u,v) € M2l\\u - v\\2 = A'. Introduce the function H(M,A')=
min
(x,u,v)£G
[T{x) + M(\\u - v\\2 - A1)} [F(x)
(26.6)
and show that each Mi > 0 is in agreement with a uniquely defined A'. From the form of function (26.6) we see that Mi minimizes fi{M,Al) when there exists the point (x,u,v) € M2t such that ||w - v\\2 - A1 = 0, and such a point in M21 represents (xl,ul,vl). If A' > 0 then there exists 1 (x,u,v) e G for which \\u - v\\2 < A and by the Kuhn-Tucker theorem for each point (x,u,v) 6 M2', (x,u,v,Mt) is the saddle point of the function Hx) 4- M{\\u - v\\2 - A') in the region G,M > 0. But for all points (x,u,v) € M2' however, Mi(\\u-v\\2-Al) = 0 must hold, and since M, > 0, then'for all these points ||u - v\\2 = A1 If A' = 0, then ||u - v\\2 = A1 also holds for all (x,u,v) e M21. Indeed if we assume that the point (x,u,v) E M2' will be obtained for which \\u-v\\2 = A1 > 0 then, iterating the preceding reasoning, we get that A' = A > 0. It is by this that we have demonstrated that a uniquely defined A' corresponds to Mt > 0.
178
Systems Optimization
Methodologg
L e m m a : The chosen {M,} corresponds to the sequence {A'} for which A —► 0. Proof: See Ref. 21. Denote by M2 the set of all solutions of prob lem (26.2). T h e o r e m 26.1: The chosen {M,} is in agreement with the sequence {x1} for which p{x',M22) — » 0 . p(x',M Proof: Denote X(A 2 ) = {x\F{x) < <(x),
||,-v||v
(x,x,v)vG},
,1 = 1 , 2 , . . . , .
The set X[Al) involves a set of all xl minimizing F{x) with the constraints ||u - vf < A' and (x,u,v) € G. Indeed, from xf'l € {x\F(x) {*|F{«) < F(xl), F(x'), \\u ||u - v\\2 < A', (*, u, v) e G} G) it follows that l xi l'eX(A e X (),A ' ) ,
(26.7)
since F{xl) < F(x) for r = 1,2,,... From A' — 0 (by the lemma) )i /-»oo
follows that X(A') — . M 2 2nd the latter, considering (7), implies l p(x',M p{x ,M2)2 ) — > 0 . i —»oo
The proof is completed. §26.2. We consider here the penalty method for the mathematical program ming problem which generates an iterative process at N levels. Moreover, calculations can be performed at N - 1 levels, and this gives the reason to call the process two-level as well. Consider the following problem of mathematical programming min{F(a;)|G < 0, min{F{x)\Gi(x) i (a;) <
1, N} ,, i» == TjV}
(26.8)
here x G En, F(x) )i a functiono Gt{xi( (j = U V ) a aector functiont Introduce the point variables Vi € En {i = 2,7V).
Decomposition
Based on Optimization
Technique
179
Denote G1={x|Gi(x)<0},
Gi {yi\Gi{ ) < 0}, Gi = = {g i\Gi(yyii)<0},
G = d
i = 2,N, 2jf,
x ••■ xG x GNj , .
Consider the problem equivalent to problem (26.8) F (ax; ) J£ £ ||x m i n iJm ||ar - Viytf ||2 == 0, 0,
(* ( sl,Sj8 i> ... , . . ,j, . , ^w )) 6 GG I ..
(26.9)
Assume that Gi is a bounded set, G is convex and closed, the function F(x) on Gj is convex and differentiable. We will set forth the penalty method for solving (26.8), using for this purpose problem (26.9). Let {Mi} be a strictly increasing sequence ef positive numbers rsenal ties) such that Mi —^ oo. Define the eroblem
M / ^ |l| la xr --3i/ /l |J| 2 (x,y (x, y22,... ,... m i n |^FF((ax; ) + M,
,yN) E e GGVI
(26.91)
and denote its solution by (xl,yl2,...,ylN). Thuss we eave reduced solving problem (26.8) to solving a sequence of problems (26.91) (/ = 1,2,...). We will solve problems (26.91) by the componentwise descent method, the iterative process of which in the present case resides in the following (1) Fix yf = y1'1 (i = 2JV) (with I = = 1°l f( = 2jf) arbitrarily) and define x1' as a solution of the problem
{
are ehosen
N
F(x) | | x\\x-yf - y f ||2 f x €l G F(x)++Mi M^l J2 e id}
)
t=2
(2) Fix x11 and define y\l (i = 2jf)
as a solution of the problems
nundl^-ftlHweGi},
(i = (i=XN). 2,JV).
(3) Fix y11 (i = p ) , define x21, etct The process of applying the method forms the sequence {xk\y\\ ....»#^} any convergent sequence of whichh
180
Systems
Optimization
Methodology
under the above assumptions, converges to one of the solutions of problem (26.21). Following the analysis carried out in §26.1, denote (26.10) t =2
and replace everywhere \\u-v\\%mN
by £ \\x- y,-||2 and (u,v\ by the point
(ft, • • •>m)- Following gheorem §26.6, we erite p{x\Ml)—►O,
(26.11)
I—► O O I—►oo
where M 1 is a set of all solutions of problem (26.8). According to the lemma from §26.1 A' —♦ 0. Therefore, considering (26.10), (26.11): 1—oo
rtSj.M1)—0
(i = 27/V). 2,N).
i—►OO I—► o o
Note that the method described when seeking the solution of the system Gi(x) <0(i = IjV) provides the solution of the problem min< ^ ]| | |xa -; -j /j /i i| |||22 {x,y Ot,fte,...,yjv) G G> &G> . . 2, ■ ■ ■ ,VN)
§27. DECOMPOSITION B A S E D ON SIMPLEX METHOD MODIFICATIONS §27.1. Consider some decomposition approaches for LP problems when implementing the simplex method which, adjusting itself to specific char acter of the problem, carries out partitioning into subproblems and thus enables one to reduce calculations. To begin with, we introduce into con sideration a new designation of matrices indicating a set of their rows and columns. Thus, the symbol A[M,N] denotes a matrix, the row indices of which run over rhe set M and dhe eolumn indices stand for the set N. In this case ^[Z,/] designates part of the matrix A[M,N] composed of its elements standing at the intersections of the rows with the index from I and the columns with the indices from J. The present paragraph deals
Decomposition Based on Optimization Technique 181 with LP problems the matrices A[M, N] of which involve a special sub-
matrix A[M0,N0] of a sufficiently large dimension. Let Mi = M \ M0, N\ = N \ NQ and in the general case we obtain the following square parti tion of the matrix A[M,N]:
If the purely special classes (Mi = Ni = 0 ) are eliminated from consid eration, then we get three classes of problems: problems with horizontal bordering (Ni = 0 ) , problems with vertical bordering (Mi = 0 ) , problems with two-side bordering (Mi / 0 , Ni ^ 0 ) . We shall not make any suppo
sitions about the number of elements of the sets Mi, Ni, though application of the following decomposition methods to LP problems with bordering is
worthwhile only when Mi, Ni is substantially less than the rank of the matrix A[M,N]. Note that the Dantzig-Wulf decomposition method described in chapter 3, §9 constitutes a simplex method modification for solving LP problems with horizontal bordering, and those with vertical bordering when the orig inal problem is replaced by the dual. §27.2. Consider general description and substantiation of one decomposi tion method based on the simplex method - the method for decomposing system. 3 Assume that a system of constraints for the LP problem is divided into two parts so that the problem takes the form minc[7V]x[AT], A[Mi,N]x[N]
= b[Mi], x[N] > 0
A[M2,N]x[N]=b[M2}. (27.3)
(27.1) (27.2) (27.3)
Denote by X(I) for any / C Mi U M 2 a set of solutions to a system of constraints with indices from I.
Suppose there is x°[N] e X(M2) and the set I0 c M2, which contains indices of all the constraints active on x°[N]. Solve the following problem
mxn{c[N]x[N), x[N] e X(I0 U Mi)
182 Systems Optimization Methodology
If X(IQ U MI) - 0, then also X(Mil)M2) = 0 with the objective function being unconstrained in (27.4) the set IQ expands (if c[TV]a;[TV] is constrained in general on X(Mi U M 2 )). Denote by x°[N] the minimum obtained, and by 70 a set of the indices active for it so that 70 C io U Mi. If x° [TV] €
X(M2 \ Io) then this is the solution of the original LP problem. Otherwise, consider the points x\[TV] = Ax°[TV]+(l-A):r0[TV], A e [0,1] and find among
them the point x\0[N] € X(M2) corresponding to the largest possible A0.
Since the point 5° [TV] € X(I0), then A0 > 0 and if only one such index io € M 2 — IQ is found that a constraint with this index is active in x\0[N}.
Denote by Ii any subset in M2 containing the indices active on x\0[N]
constraints, and take as xx[N] any point in X(M 2) the active constraints of which are in Jj. For instance, one can take a^fTV] = £;\0[TV], and collect in I\ the indices of its active constraints. The choice, however, can also
be made otherwise. In any case, the constraints with indices from Jo n -Jo are active at the point x\0[N] and, therefore, Io C Mi U I\. Moreover, the point a:0 [TV] does not obey the constraint with the index io 6 I\. Again solve problem (27.4), this time on the set X(Ii U Mi). If the set is empty, then so is the set X(M\ U M 2 ). Otherwise, the form c[TV]x[TV] achieves a minimum on X(Ii U Mi) since it achieves the minimum at the point i°[A^]
on the set X(I0), and I0 C h U Mx. Denote by xl[N] the point of the minimum obtained.
Since f 1[7V] obeys the constraint with the i0 index, and x\N] does not, then it implies that x°[N] ^ ^[N]. And since the constraint set, with which the point ^[./V] is obtained, involves all the active constraints of the point
x°[N] then c^x^N] > c[N]x°[N]. For the pair of points {x^N^x^N]) ita is possible to iterate all constructions obtaining the points x2 [TV], x2 [TV], etc. Thus, we get the sequence of sets Ik C M 2 , k = 0 , 1 , . . . and the sequence
xk[N], k = 0,1,... of minimum points c[TV]x[TV] on the set X{I k U Mi) for which c[N]xk+1[N] >c[N]xk[N], fc = 0 , l , . . . And since the number of ctTVjj^TV] is uniquely determined by the set Ik, then the sequence xk[N],k = 0 , 1 , . . . is finite. Therefore, for the finite number of steps we will either obtain the solution of the original problem or establish its unsolvability. §27.3. We consider application of the method described to the problem in canonic form.
Decomposition Based on Optimization Technique 183 Now, assume that problem (27.1) - (27.3) is specified. We take system (27.2) as the second group of constraints. The issue of selecting the initial
point x°[N] can be solved in many ways. If form (27.1) is bounded below on a set of solutions of system (27.2), then one can take its minimum point as x° [N]. In a more general case one can proceed from the fact that there is a known row j/i [M2] with which (27.5)
(c[N] -yi[M2}A[M2,N})x[N] (27.5)
is bounded below on the solutions x[N] of system (27.2). Let x°[N] be a minimum point of (27.5), and J 0 its associated basic set (J 0 C N). If we denote by yi[Mi] the solution of the problem dual with respect to (27.5), (27.2), then the row y\[M\ U M2] satisfies the conditions
xy [Mj U M2]A[M2 UMUN\J0}< c[N \ J0],
(27.6) (27.6)
j/i[Mi U M2]A[M1 u M2, J0] = c[J0]. Because of this, either the system x[N\ J0] > 0, A[Mi, N)x[N] = b[M{\, (27.7) A[M2,N}x[N}
(27.7)
= b[M2]
is inconsistent and, therefore, so is system (27.2) - (27.3) or attains the minimum on the set of solutions (27.7) - (27.8). Now, obtain the minimum point of (27.1) with the (27.7) - (27.8). Since the matrix A [Mi, Jo] is nondegenerate, known x[Jo] are not related by the non-negativity condition, (27.7), we write
(27.8) form (27.1) constraints and the un then, using
a:[Jo] = 7J[J 0,Mi](6[Mi] - A[MUN\ J0]x[N \ J0}), (27.9)
(27.9)
where D[J0,Mi] = {A[MU Jo])'1. Substituting (27.9) in (27.1), (27.8), we get the problem
miac[N\J0]x[N-Jo], x[N \ Jo] > 0,
(27.10)
A[M2,N\ J0]x[N \ Jo] = b[M2] - A[M2, J0]x°[Jo], (27.11) (27.11)
184 Systems Optimization Methodology where the constant summand is omitted in (27.10) and
c[N\J0}x[N\J0} = (c[N\J0}-y1[M1}A[Mi,N\J0})x[N\J0}, A[M2,N\Jo}=A[M2,N\J0)-A[M2,J0]D[J0,M1]A[Mt,N\J0].
(27.12)
Evidently, the virtual calculation of all the matrix columns (27.12) can be justified only with the block diagonal matrix A[Mi, N] of a small dimension, though. Problem (27.10) - (27.11) can also be solved in the following way. Sup pose for this problem there exists the next basic set K C N \ Jo, the columns of the basic matrix J4[M2, if] are obtained or the matrix D[K, 2] M2] inverse to it is known. Obtain the row y[M2] = c[K]D[K, M2] and test the optimality condition y[M2]A[M2,N\ JQ] < c[N \ J0]. (27.13)
(27.13)
If we first calculate the row y[Mx] = yi[Mx] - y[M2]A[M2, J0\D[J0,MX], (27.14)
(27.14)
then (27.13) will be written as y[M2 U MX)A[M2 U ML TV \ J0] < c{N \ J0]. (27.15) (27.15)
We need not know the matrix D[Jo, M\\ since its utilization can be replaced by solving the system w[M l]A[M1,J0] - y[M2}A[M2, J0].
Then y[Mx\ = y\[M0] - w[M0]- Obtain the number j' e N\ J0, on which
(27.15) is violated, and calculate c[j'] = c[j'] -yi[Mi]A[Mi,j'}, and then solve the system
A[M1,J0]g[J0]=A[M1,j'}
and obtain the column A[M2, j'] = A[M2,j'] - A[MX, Jo]g[Jo], to be intro duced into the basis. Suppose the optimal solution x°[./V\ J0] of problem (27.10), (27.11) has been obtained. Substituting it in the right-hand side of (27.9) or solving the system
A[Mi, Jo}x[J0] = b[Mx) - A[MU N \ J0]x°[N \ J0]
Decomposition
Based on Optimization
Technique
185
we get the remaining components of the minimum point x°[N] of form (27.1) on a set of solutions to (27.7) - (27.8). If £°[J 0 ] > 0, then the column of x°[N] constitutes the optimal solution of problem (27.1) - (27.3). Otherwise, we obtain
Ao A 0== m m ii nn j
o r M ^J - o f T 3; €€J oJo, zf ° [mj ]<<0o l {x°mi°[j]' ' ° }\ y [3 \ [) \ x
x
(27.16)
and set xXo[N] = \0x°[N} + (1 - \0)x°[N\. If J 0 is the basic set with which solving problem (27.10) - (27.11) terminates, j 0 is the number transferring a minimum to (27.16), then the numbers of all positive components are in the set Jo U (Jo \ jo)- They can be taken as the initial for the next step, though the set may not be basic for (27.2). It can be remedied in the following way. We solve the following LP problem min^xlj], m in^xtj], rr ' % , X[J0 U J 0 ] > 0 , x[J0 U J 0 ] > 0 , A[M J0]x[Jo 0 U J 0 ] = b[M b[M00]. A[MXX,, J 0 U Jo]z[J
(27.17)
As an initial feasible set for (27.17), we can take Jo and, after solving it, we will obtain a basic set Jx and a feasible solution of x^[JQ U Jo]- Since xXo[J0 U Jo] satisfies the constraints of (27.17) and the functional for it is equal to zero, then a % ] = -. For the set J J i b t a i n e d dnd dJ[iV] we cac carry out the following step of the method. Evidently, the methods for solving LP problems with horizontal bordering can also be applied to the problems with vertical bordering. For this purpose, it suffices to turn to the dual problem. Moreover, these methods can be applied for problems with two-side bordering, cutting off first a horizontal bordering part, as described, and considering the remaining problem with vertical bordering as the special. In turn it is possible to use the decomposition method each time when solving the special problem. §27.4. Consider one more decomposition scheme based on the simplex method for the general LP problem 15 max{cx\Ar
= 6, b, x > 0 } , =
(27.18)
186
Systems
Optimization
Methodologg
where A = A[M,N], M = { 1 , . . . , m } , AT = { 1 , , . . , » } , m < n. Partition the set M into the subset Mk so that M = (J Mk, Mk n M, = 0, Jb ^ j and isolate the submatrices A[Mfc, N] of the matrix A[AT, N}. Assume that the blocks A[Mk,N] are loosely yound, d.e. when solving (27.18) by the simplex method all the basic matrices involve a considerable number of the columns A[M,j] among which onllyne from m[Mk,j] is different from the zero one. At each iteration of the simplex method we need to solve two systems of the equations y'[M}A[M,R}=c'{R], A[M,R]q[R) = A[M,j], A[M,R]q[R]
(27.19)
j£N\R, jeN\R,
where A[M, R] is a basic matrix, R = {ji,..., jm} C N is a subset of basic numbers. We will employ the above assumption to transform A[M,R], By rear ranging the columns (if it is necessary) we can transform it to the form /AiM^P] /A[M UP\
A[Mt,Qt]
A[M,R]=
... ...
0
\ (27.20)
■ ■ \A[Mt,P]
0
...
A[Mt,Qt]J
where the matrices A[MS,P], A[MS,Q3], S = T~t tan also be nonzero, though not both at once. Since the matrix A[M, R] is not expressed, then the Binet-Cauchf formula implies the existence of the partition P = (j Ps such that the matrices A[Ms,ks] = (A[MS,PSs],A[M A[M„k.) },A[MSS,QSS}),
s= = 1,* M
are square nonsingular (jfc, = Ps U Qs). We form the block diagonal matrix fAlMukx] l
D[M,R}=
V and consider the product
0
J
...
...
D-1[M,R)A\M,R].
0
\
A[Mt,kt}J
Decomposition
Based on Optimization
Technique
187
It is evident that )
^-Mfcs,M^[Ms,P,]=(^ )), ^ -'[w.m,p.i=(^),
A i%
Q
,-| l „M,w«„ej = (; p ;; g ft ;).
- >*^ H%:liY
Isolating in each block of the matrix the upper Ps rows and the first P columns, we get the square nonsingular matrix [P11,M ,M11]A[M ]A[Mil,P]\ ,P}\ /A-l1[P G[P,P]=\ \A-l{Pt,Mt}A[Mt,P}J A[ks,j} ,j) = A-'[ks,Ms)A\Ms,j}, (27.21) y'[ks}=y'[Ms]A[Ms,k }. y'[ks}=y'[Ms]A[Ms,kss]. With the aid of A[PSJ], q[Ps], y[Ps), c[Ps], s = M form the vectors A[Pj},q[P],y[P},c[P}, A[P,j],q[P],y[P],c[P],
(27.22)
where P = \J Ps. We employ now the matrix D[M,R] to transform equa tions (27.19)=1 y'[M]D[M, RjDR]D~1l [R, M]A[M, R] = c'[R], 1 D-1[R,M}A[M,R}q[R} D-1 [R, M]A[M, R]q[R]==D-D'1[R,M}A[M,j}, [R, M}A[M, j},
(27.23)
(27.23)
Considering (27.21), (27.22), the definition of the matrix G[P,P] and the structure D[M,R], A[M,R], obtain from (27.23) a's[P] [P) =
l c'[Qs]Ac'[Q [Qs,Ms}A[Ms,P}, ,P), s]A-\Q t
y'[P] = (c'[P]-^^[P])G- 1 [P,P], 3= 1
y'[Ks)] = (y'[P {y'[P }), s},c'[Q sld[Qss}), 11
y'[Ms}=y'[k ]=y'[ks}A~ }A- [k [ks,M s,Ms], s}, q[P}=Gq[P] = l[P,P}A[Pj], G-1[P,P]A[Pj], 1l q[Q,] q[Q [Q„M.](A[M„j] 3] = A- [Q s,Ms}{A\Ms,3}-
- A[M A[M„P]q[P)). s,P]q[P}).
(27.24)
188
Systems
Optimization
Methodology
Relations (27.24) show the possibility of decomposition when solving (27.19). At each iteration of the simplex method a change is made in the matrices A-l[k3,Ma], G-l[P,P] Thus, the main part of the simplex method calculations is carried out, using (27.24) as the base, while all the other operations are performed in the ordinary way. Remark: If a particular block has a special structure, then it can be completely included in the algorithm. §27.5. Consider the algorithm for solving the large problems of linear dynamic programming which represents a further development of the above method. 16 The linear dynamic programming problem in its general statement takes the form
(
N+l
N
\
(27.25) d u
xk + yyic ] kc/tifc + Vy2 ] dkkukk ,, xk+i +k=0 Akxk + Bkk=0 uk = fk /, k=0
k=0
(27.25)
(27.26)
/
(27.27) Aixk+Bluk=fl, k = 0,N. Xk+i + Akxk + Bkuk = fk , (27.26) A\xk+Bluk=H k = 0J*. (27.27) We suppose, without restricting generality, that all the vectors of the states xk, k = 1,N are free variables (otherwise it will require only introducing auxiliary variables and equations into (27.27). Next we assume that a nonnegativity condition is imposed on the vectors xo, XJV+I, uk, k = 0, N. Problem (27.25) - (27.27) can be regarded as an LP problem and solved by the simplex method, at each iteration of which we need to solve two adjoint systems of equations of form (27.19). Evidently, the constraint IE B A 0 \ matrix of LP problemN takesN the form 0 BN AN IE BN AN 0 \
0 A = A =
E BN AN E
BN-I Rl
AN-I ji
Rl
ji
A Bpt-i °N-l AN-I N-l
°N-l
\\ 00
A
N-l
flg flg
E E
B B00 A A00 Al) Al)
Decomposition
Based on Optimization
Technique
189
which implies that the basic matrix of the problem at each iteration is as follows IE
BN R11 n N
A0 =
AN A11 A
N
E E
\ B^-i BN-\
**N-i
A0 =
**N-I
AN-I AN-\
AN_X
A.N_X
E
V \
B0
Ao A0
Bl Bl Al)
where Bk (k = 0JV)) A0, B\ (k = QJf), i j , E are the submatrices com p o s e d ^ the columns of appropriate matrices. Since the variables xk, k = 1, JV are free, then they are simultaneously basic at each iteration. We rewrite the principal equations of the simplex method (27.19) as A0[M,k]q[k] [M,k}q[k] = bj[M], [M]:
y'[M]A0[M,k]
= c'\k], c'[k],
(27.28)
where M, k are the appropriate sets of the basic matrix rows and columns, and the vector 6J[M] has zero components everywhere with the exception of the rows pertaining to the j t h moment of time. If we introduce decomN
position of the set M = \J Mk kccording to the moments of time, then
( °^ b'\M\=
b>[Mj]
and, moreover, we partition the sets Mk into Mko, Mkl according to the constraints (27.26), (27.27). At first we will consider the principal equations of the method described in §27.4 for the case of two-strip partition of major equations. The ordered series of such partitions underlies the method suggested for solving linear dynamic programming problems. Note the first strip corresponding to zero reading, i.e. isolate the rows of the set M 0 in the matrix A 0 , and transfer N
the rows of the set M 1 = \J Mk to the second strip. Take as connecting fe=i
190
Systems
Optimization
Methodology
the components of the vector Xl which correspond, in the first strip, to the coefficient matrix
--G) and, in the second, to /0 Jl=
\
0 A,
\A\J Following the method described in §27.4, we will introduce the square nonsingular matrices DQ, A: characterized by the A 0 columns multiplication coefficients in appropriate atrips (i.e. products of D^1 and A " 1 by corresponding local columns) form unit vectors. Naturally, the matrices Do, Ai are not uniquely determined by these conditions, but later on it will be shown how to make a better choice. Replace equations (27.28) with the equivalent H^lA0q
= H^b^ H^lb*,
where
1 y'H y'HoH^Ao = c', c', 0Ho &o
(27.29)
M?:) *-(? i)
Considering the definitions of Ai, D0 we have the columns of the matrix H-lA0, with the exception of those pertaining to the coefficients of the vector Xi, will become unit vectors. Isolate the submatrix Gi of the matrix HQ1AQ formed by the elements standing at the intersection of the above columns and rows which do not contain other nonzero elements. Let kl be the set of columns of the matrix A l t kQ - the set of columns of the matrix D0 and let k = fc1 Uk0. Introduce the matrices E[XX, fc1], E[Xlt k0] composed of rows-unit vectors with units being at the places corresponding to the rows A}J\ D^1 J0, which form the matrix Gu i.e.
(ElXuk'jA^J^ [ElXuk^A-'j^ Gl=
{E[XtMn^jJ-
Decomposition
Based on Optimization
Technique Technique
191
Somewhat changing the notations, we employ formulae (27.24) (E[XUuk*]A? #]A?
b>[Mi]\ bi[Mi}\
V £[*I,MV V£[Xi,fcoPo
b>[M &J[M>]/ 0]J
1 q[k # o0 \\ X!] X!] = = E[k E[k00 \\ X^kojDv X^kojDv1
(b*[M (b*[M00]] -- JJ00q[Xi]) q[Xi]) ,,
1 1 1 1 qlk qlk1 \X \XXX]] = = E{k E{k1\X \X11,k ,k1]A^ ]A^1
1 (VIM (V\MX]] -- J'[Xi]) fqlXi}),,
a' = ElX^kolc'lkojDElX1,k0}c'lk10}D-J1J0 0,, a'00lX {X1x}} =
(27.30)
1 11 1l a' J, a'11lX [X11}=ElX }=E[X11,k,k1}c'lk }c'[k1]A1]A- J ,
y' y'11lX [Xll}] = =
(c'lX lX11}-a[[X }-a[lX11})G^\ })G;\ (c'[X11}-a> }-a'00[X 1 c'lk00]}Dc'[k ]}D-x,,
y'lMo] y'[M0] = = miXi) m[X{\
tf[Xi\) E[X E[Xltukc] -- tf[Xi\) k0]
+ +
y'lM1} = MIXi] M[Xi]
-C'IX^EIX,,^}
+ c'[fc1]}Ar1 ,
where Efk0 \ Xuk0] are submatrices of the unit matrix E[k0,k0], as sup plementary (possibly after rearranging rows) to E[XukQ}, ],e remainini matrices and vectors are formed in the same way. Formulae (27.30) show the possibility of decomposition at one step. But the main difficulties of solution are associated with recurrence of the matrix Ai which has the same structure as the entire basic matrix Ao- At the same time, the analysis of equations indicates that the same operations are carried out with Ai as with the matrix Ao- This enables us to construct the recurrent process of solution, the construction of which is characterized by the fact that we can keep the structure of the remaining part unchanged by separating successively the bordering associated with the moments of time. With the basis changing, however, this feature causes us to make efforts to maintain the structure. We describe the recurrent process of solving (27.29). Similarly to the definition of the matrix
192
Systems
Optimization
Methodology
case the set of columns of the matrices D{ is denoted by fo and ki+1 Also denote l b[ki] = D-D-H[Mi]. b[Mi\. i[k]
i+1
(27.31)
Due to the basic matrix definition and the fact that the basis can involve only the vectors in b[M] possessing nonzero elements only at one of the moments of time, we get for a particular s b[ki} = 0, 5[**1
i^s. i^».
The direct move of the algorithm for solving equations (27.29) has the form qN[XN+i] l[ki] = b[ki]
=
E[XN+i,kN]b[kN], b[ki\-4qi[Xi+1], b[ki]-Jlq>[X
V^P^-i] y'N[MNN}\=y'=N[M y'N[M N\ N} =
by*-!])'
c'[k c'\kN)D}DjNl1\,
,
= m[Xi]
(27.32) (27.32)
(c'lXi)-
i = N-
+
c'lk^jD-},, c'lk^^D-},,
!N-l,...,l. ,...,!.
Prom this we get g 0 [^i] =
+ l,
Decomposition
Based on Optimization
Technique
193
The inverse move of the algorithm is performed with the formulae ,' [[ X X-,1
= =' ', [. *| A«-1i-+G, |r -«^(1£(I£xK ,++i „, t ti .| RD " (^)«)' ^^?|XJ).
+ l l
q[vt] = l[vi) - E[viyu ki]DkijD'1x
|
q[Xi] - E[vh Xl+l]q[Xl+1],
(27.33)
i= = QJf, 0,N, i[ki] = c'[h] + tf[Xi] - c'[Xt}) E[Xh ki], h}, a'1[Xl+1}=-,l[kz}D-1r0, y'[X1+1] = {c'[Xi+1] - a'x[X1+1] -
11 +i[k , + l)}D; 1'[k l)}D;
T l,N. jv.
Thus, (27.31), (27.32) prescribe a sequence of matrix operations to solve the basic equations of the simplex method. The remaining calculations are carried out in the regular way. From formulae (27.31), (27.32) it is seen that in the process of solving the basic equations one should maintain a comparatively small amount of information apart from going from one iteration to another {D-\G~l)v vi.thev vectorg ql[Xl+i], a'l[X1], i = Tjf.
Chapter
7
DECOMPOSITION AND
AGGREGATION
The present chapter deals with the decomposition methods based on introducing aggregative variables (macrovariables). Aggregative variables are frequently introduced as sums of some components of initial unknown values. In other cases the weighted sums are taken. Thus, each time we have an intermediate problem in the aggregative variable case, which involves a smaller number of unknowns as compared to their quantity in the original problem. The aggregative problem is either solved in a finite form or acts as the coordination in the iterative process. An exact or approximate solution of the original problem is conclusively constructed according to some rules. §28. AGGREGATION METHOD FOR SOLVING A SYSTEM OF LINEAR EQUATIONS §28.1. Consider the following system of linear equations (28.1)
x = Ax + b, b,
where x, b are n-dimensional vectors, A is (n x n) matrix possessing the properties: n
0
ij
€ {1,...,«}, k>0,
J^oySl,
!i e£ { l! , . . . , n« } . 194
;G{l,...,n>,
Decomposition and Aggregation
195
Let x = (xi,.., ,xn) be the solution of system (28.1). By adding all the equations of system (28.1), obtain the scalar equation X = aX + b, where
n n
X=xx-{
\-xn,
a= ^atpi;
pi = -^,
t=i
(28-2) (28.2)
n
6b = = bi 6i + +
(-6,1, h 6n,
aj aj == }7yy
aajj. jj.
i=i i=i And conversely, if the X solution of equation (28.2) is known, then
Xl = ai v xi=(j2a iiPj)x
( HI i i }X ++bib-i.
(28.3)
It is evident that problem (28.1) will be solved if we succeed in defining the Pl numbers so that
=
T
* T
nn
i=l
We will set forth the iterative method for solving the system of equations (28.1) described in Refs. 43, 44. Let x° = ( a ? , , . . ,i°n) be specified as an initial approximation of solving (28.1). Let n v-(O) _ V ^ 0 A — Z^^ii i=i i=\
(0) _ Vi
(0) ^^ i i_ _ _ ^-(o) ■
Consider X 1 according to the following rule: X( 1 ) = xW = (a1lPp[(10)0) + + --- + a n P W anpW)XM+b, )X(1)+6, or, denoting obtain
aa ( ( oo )) == aa ii pp (( oo )) ++
.. .. .. ++
aa nn pp (( o o )) ii
xM=a0X^ + bb X^=a°XW
(28.4) (28.4)
196
Systems
Optimization
Methodology
and further (N+l) = X (jv+i)
X
+ 1 )+ J , Jw )^ +{ ND = I^£^a jW +b, P$. \x
(28.5) (28.5)
i) > +i) xr +^ *f}==lw (w+i) =(E^# _ J2 a p )^ X + hi, pf -^y■ /
n
\
[ i]P
x
(N)
{N+1)
]
(28.6)
(28.6)
From (28.5), (28.6) we get (N + l) Pi
n
n
aa%V P
~ - Xjjf(N+l) (N+1)
Z^ ^ J3 3 3= 1
^^rr*2^ *2^PPllPPJ3 3= 1
(28.7)
n a = ^2( n ij i=i
+ ^/?J)PJ;V);
3= 1
k rTi j = = j—;;
O
(28.8) (28.8)
1
0j Pj = - I l-aj. - dj
Introduce the matrix S = {«»}, {stJ}, Sij ^ ==aij, +r+i0j,^ , « £ i,je{ | l
■>}
and rewrite (8) in the form pC+D p ( (o0)) • p( w +D == 55p pWw == sS( /^v ++ i1)V
(28.9)
The matrix 5 has the properties: (1) &ij stj > 0, n n
(2) ^ 1=1
i , j = ll,n, ,n, n n
sSyy = Y?,( ^ ( aai3t J + a{ + Pj = 1. 1, Jj = T~n. + Tri/3,-) iPi) ~=ai+ Mli=l
Assume that the matrix S is undecomposable, i.e. there is no such permu tation matrix P* that
pFr>*^McfpFr>*\~l *) ) - =_ [ 5gJl | |g2 ' -[
0 | s, .
Decomposition
and Aggregation
197
where Si is a square submatrix. A sufficient condition for this is nondecomposability of the matrix A. It is known1 that (1) the sequencceS N N converges to t o a t r i x T; (2) the column vectors of the matrix T are alike and equal to the vector t>0; (3) for any vector p (pi > 0, £ p t = 1) i N lim 5 p p t, SNp
t= = ST
From (28.9) and the above properties of the matrix S it follows that p{N+1) p{N+i)
X (N+i) =
x(»+v =
v p
N — oo
b—-
1N
Erf -~
>x >x
ji
and
(N+l) X\ '' X\
>X >Xtt..
N N— — oo oo
Thus the convergence of the iteration process has been shown. From the calculation standpoint, using the obtained results for con structing the iteration process is more handy. Definition: Xk is called Jk aggregative variable if Afc = — y
Xi, Xi,
k = — 1,1; 1,1;
ieJi. i€Jk
Ji n Jj = o, 0,
ji /± ji ;
Ii
(J JJ<<=={l,2,...,n} {l,2,...,n} The matrix A(nxn)) with huch hggregation, is convoluted into the matrix A(lxl) the entries of which arrebtained by transforming the A matrix ele ments as follows: a-ij = = ^2a ; p«; Oij: ^ a sts+p t t+
=
s« 6e^Ji, .> e
ft Pt = I T '
t e€ JJy, < ,;
=1
i,j i,jf = = 1,2, . ..,/,
5J» -
2^ ft = * •
(28.10)
198
Systems Optimization
Methodology
Denote b b bt=J2 s: s i h = 2J
,...,bi)TT(b(b11,...,bl)
B B=
s€Ji
Then system (28.1) transforms into the system X = AX + B.B.
(28.11)
When constructing the iteration process we specify, as before, the initial value of f{°\ 5 =T~n. At the first step we eet the values of Xu..., Xt by solving the system (28.12) X^ = A^X^ + B.B. The entries of the matrix A^
are derived from (28.10) if we set (0)
_ x\t ' ~lAo)' A
Pt
t<EJ
j
i'
teJ,
At the same time, the knowledge of the exact solution of (28.12) is basically not necessary. Solve (28.1) as follows: Define
0) ,(o) J2xi= y(o), =Y^, = £4°> y(o) = fc=i
,,
x^
0) ±±_. ?iJO) =_ fy(0) k ;' fc v
4 0) = Ii =Xl 0 ) ;
—
/ V^ -(0) (o)
-(o)
5 i== l J > i=l
i=i
Let Y^
(28.13) (28-13)
be obtained from solving the scalar equation
Instead of denning Xk
3=1 y(i)
i=\
= a
(0)
y
(l)
+ 5
(28.14)
y(D=a(0)y(l)+£ from (28.12), we set
(2gl4)
X f =Y^ \ , k* = x ^[ V = TjM
(28.15)
{ l) k
q
0
Decomposition and Aggregation
199
and next obtain Ii
-l^EE^^H, -Mj=i
teJ,
(28.16) (28.16)
Similaxly, the value of xs at the (JV + 1) step of iteration is obtained from the formula:
j=i
tejj
p[N) = x[N] IX\N\ N+i)=Y(N i)q{N^ AxiK -+X
J
'q"
y ( N + 1 ) = a ( N ) y ( N + l)
t£jj, ,
+ 6i
J
(28.17)
T7lj ~ J=1,1,
qW =
±JL__
We will call the process of solving system (28.1) in the first case the singlestep iteration process, and the process described just now the two-step iteration process since in the former case all the variables of system (28.1) are aggregated into one variable X, whereas in the latter a similar aggregation process was accomplished for two steps. As a result of iteration with the number TV of the two-step process we obtain the values of the variables xs = x\ Let xs — x\ be the value of xs after the iVth iteration in the single-step process. We will demonstrate convergence of the two-step process. The initial values for both processes are the same, i.e. p°t=ftq°,
J = 1J 1J
(28.18)
= ££,(o)g(o)
(2gi9) (28.19)
t£jj,
Compare (28.3) with (28.14), obtain I /
n n
i=l i=l
j=l j=l
5
3=1 3=\
tp( ]
i = l jj=l =l i=l n 71
= Yl as ° = J2 a^pf] = a° s€UJ, teuj,
i,j = l
200
Systems
Optimization
Methodology
Prom (28.3), (28.14), (28.18), (28.19) we get X(D=r(i). X^ = Y™ .
(28.20)
By using (28.6), (28.13), (28.16), (28.18) and the fact that J3 do not inter sect, we obtain
j= \ 3-1
t£jj t£jj
= Y: W - bsb.. . £ aJ^JCW astP^X^ = X«?> teujj
Similarly, bb induction we prove that x[Nx = x[\). Naturally, the process of solving (28.1) can be represented not only as the two-step but also as the multi-step procedure. The method described enables one to solve (28.1) numerically at the cost of solving the problems each of which accommodates a little more than part of the matrix A or a certain convolution A of the matrix A. §20.2. The iterative aggregation method described is particularly efficient if the matrix A in (28.1) has the block diagonal structure
A _
/on
0
...
0 \
j «21
»22 0-22
■■■ •■•
00 \ \
\o a mm i
am m 2
...
o 0-ram m m I/
All the variables of xu (* = hm), i € Jk are assume to be aggregated into one magnitude of Xk (k = l , m ) . Thus (28.1) is written in the form: Xi
=a11X1+b1,
X22=a2iX = 1+
a2lX022X2 +2+b 6 22, x+a22X
cimiXi + am2X2 A Xm = a-miXi -I
+ ammXm + bm .
We set that at each step of the iterative process (28.1) is uncoupled into m equations to be solved separately.
Decomposition
and Aggregation
201 201
§29. A G G R E G A T I O N M E T H O D FOR T H E BLOCK P R O B L E M OF LINEAR P R O G R A M M I N G §29.1. Consider the following LP problem (23) c 0 x 0 + ^ cCiXi + J^^ dij/i dit/iI\,, min I coio ^ + I i=l t=l J fc fc 400X° a;0 + + ^ ^ lAiaXi; t = =& 6o, 0, JA
(29.1)
i=l
B{Xi + + Diyi = bi, x0 XQ
> 0;
Xi > 0,
J/J yl
i= = l,fe, l,k, > > 0,
j = 1, fc.
x tt > 0,
y *t > > 00
Assume that the conditions B 5 ttxxz t + +D Dan k, lVl = b h
prescribe for a l l i € { 1 , . , . , N} the constrained polyhedrons. Aggregation will be taken as the following change of variables xi=a Xi = c*iXi, iXi>
i=TJ, i = 1, fc,
(29.2)
where a = ( a l , . . , a a)ist the vector of aggregation weights, a, ^( = ljfc) a particular fixed vector. We wiillay that aggregation is feasible if a, a t fo0 all r = 1,=1 By substituting (29.2) in (29.1), we get the problem with aggregative variables min I< cox c 0 x00 ++^^ c~iXi CiJ¥"» + ^ d^yi diyi > , I i=l i=l J fc
A 0x0 + ^4oXo + YiaiXi=bo, 2_j Q-iXi = &o i
(29.3)
i=l
biX% + Diyi = bi, x0>0;
Xt>0,
i = \,k,
3/i>0,
zz=I7fc, = l,fc,
where aSii = A v4iQ:i, iah
6i Bi«i, k = Bm,
Ci Ci —= CiOti. cta%.
(29.4) (29.4)
202
Systems
Optimization
Methodology
Since the aggregation weights can be selected in various ways, then one original problem corresponds to an infinite number of different aggregative problems. We then establish some solvability properties of these problems. Theorem 29.1: If the original problem is solvable, the aggregative problem has a feasible solution and aggregation is feasible, then the linear form of the aggregative problem is bounded below on the entire set of its programs. Proof: Let {x0, Xl: yt} be a feasible solution of aggregative prob lem (29.13). Set ii = arXi. Evidently, {x0,xi,y1} is feasible in the original problem, and the value of its functional on it Cp coincides with the functional value of problem (29.3) on ththet {
I
A0x0 = b0,
Dlyl=bl,
£i
x0 > 0,
(29.5)
I
y, > 0,
(29.5)
i=
l,k.
T h e o r e m 29.2: So that any one of the aggregative problems will have a feasible solution, it is necessary and sufficient that the added problem might have optimal solutions. Proof: Let {x0,§i} be the optimal solution of (29.5). By adding to it Xt, i = l,k, we get the feasible solution of any aggregative problem. If the added problem is unsolvable then so is the aggregative problem obtained with Q = 0. Thus, if the original and the added problem are solvable and aggregation is feasible, then each of the aggregative problems is solvable, which will be further kept in mind. Let {xo,Xi,yt} be the optimal solution of (29.5) with a particular vector a the functional value being 0 Assume that x% I QlXl and c a l l l t h e e t { * , * , * } a disaggregated solution. It is feasible in the original problem. Thus, if {ipa} is a set of optimal values of the functions of aggregative problems, then we have
(29.6)
Decomposition
203
and Aggregation
Theorem 29.3: Let au a2 be the aggregation weight vectors of two ag gregative problems, here <*i = ka2 and k > 0, then the sets of their disag gregated solutions coincide. Proof: See Ref. 23. We will establish now the existence of aggregative problems the func tional values of which coincide with the functional optimal value of original problem (29.1). Let {x*0,x*,y*} be the optimal solution of (29.1). Suppose that en = x*, Xi = 1, * - L i , then we have the eet tx*,j,3/*} being fea sible in the corresponding aggregative problem which gives the functional value v* Thus we have the equality min<^ min (faa = tp*
(29.7)
a
The aggregation vectors a, with which equality (29.7) is achieved, will be termed the optimal aggregation vectors, and their associated aggregative problems the problems with the optimal value of parameters. Aggregation, which leads to the problem with optimal parameters, will be called optimal. The next purpose is to construct the iterative process which, starting with a feasible aggregation, leads to optimal aggregation. To formulate the basic theorem we have to introduce additional defini tions. Let {x0,Xi,yi} be the opttmal lolution nf the eggregative problem with the aggregation weighi vector a. Setting b0 = b0 = J2
tti
we
can
i=l
construct the LP problem min m i n CQXQ CQXQ , ,
A0x0=b A0Xo=b x0>0x0>0. 0, 0, It is evifient that r is one of the optimal solutions of the problem We will identify the problem connected. Let 6b be an arbitrary vector of the same dimensionality as b0. Introduce the so-called connected parametric problem minCQXQ , mincozo, (29.8) A0x0 -b + 06b, x x0>0. >0. 0 A0x0 =b + /36b, 0
Q
With 0 = 0 the solution of (29.8), as we eee, existst and dith h > > it may not already exist. We need to fulfill the condition. Condition 29.1: For any vector 6b and a particular /? > 0 the solution of (29.8) exists sifi exists sor 0 3 0. For problem (29.8) we formulate the
204
Systems Optimization Methodologg
dual with 0 = 0.
maxM,
u'Ao > c°. Denote by v0 the constrained part of the optimal program estimate vector set of this problem, i.e. v° = {u0\u'0Ao A0 < cQ,
(c0,x0) = (bo,u (b0,u00),),
\\u0\\ < 00}
For any uQ € v° we can form the vectors ar = uoAi ai u0Ai — - ad . .
(29.9)
Consider rhe local problems with each i € h i , . . . , it}: max^x, maxiffiH + ddivA, %yi}, Si^i + Diyi A j / , = b,, BiXi bi, *i > 0, Xi>0,
(29.10)
yi yt > 0.
Let {x;,3/i} (i = TJc) be some optimal solutions of problems (29.10), and z{ = aiii + diyi be ehe ffnctional opttmal value of fhe ith ppoblem. Establish solvability of the problem. The elements of the disaggregated solution {xl,yi} are feasible in the corresponding local problems. Therefore, if Zi = axi + d{yx, then the inequality z; < % is fulfilled or, by introducing k
k
the designations z = £ zit z = £ zh the inequality z < z is completed. T h e o r e m 29.4: (Aggregation optimality criteria). If condition 29.1 is met, then for aggregation optimality it is necessary and sufficient that for a parttcular u0 Q v0 the following equality might be fulfilled 1=5. (29.11) 1=5. Proof: Sufficiency: Consider the problem dual to the aggregative max < u0b0 + ^ P u^i Uibi > , u0A0 < cc00, uocii u^i <
UiDi < di,
i = 1,k.
(29.12)
Decomposition
and Aggregation
205
Let {x0,Xi,yi}, {u0,Ui} be the optimal solutionnsf (29.3)3 (2(.12), resrec tively. By the second duality theorem we have Xl(uo,al-u1bi-ci)^0. By utilizing (29.9), (29.4), we rewrite the above equality as (ffi,*<)-JEi(flit5i)
= o1
where x{ is the element of the disaggregated solution. Moreover, btXl = hi - Di§i and, therefore, %ibi Uib
(29.13) (29.13)
If (p ii sth eunctional lptimal lalue eo fth eagregative eroblem, ,ten ni ti simultaneously the functional value of the original problem on the disag gregated solution which is feasible in this problem. Prom the optimality in the aggregative problem and (29.13) we get k ft
(29.14) (29.14)
t=i
Consider now the problems dual to the local (29.10) with i = 1 , . . . , k u&nvibi, tmnviki, vtBt < (Ti, (Ti, ViBi
Dll < dxt. vllD
(29.15)
Let vu i = T7k be the optimal solutions of those problems, i.e. z, = v^. Then the set {u0,i;t}, (i = ITS) is admissible to the problem being dual to the original. This can be managed if we develop the first inequality in (29.15) with regard to a% defininionn , t h eunctional lalue eo fth eroblem dual to the original being equal kit =u 6e = u0bbo Y^v^Vibi obo= u0b0 + z-2. 0 + J2
If 3 = z, then (29.14), (29.16) imply
(29.16)
206
Systems
Optimization
Methodology
i.e. the functional values of the original and the dual problem coincide on two feasible solutions {x0,Xi, y~i}, {uo, *>«}• This signifies optimality of the latter in appropriate problems. Thus, we conclude about the optimality of integration when (29.11) is valid. Necessity: Consider formation of local problems with the optimal solution of the dual problem for the connected (29.18). Let «o be the optimal program estimates corresponding to the basis XQ. Form the local problems and let z < z. This means that for a particular k' < k the inequalities zz < £i, i — l,k' are fulfilled. Furthermore, let {xi,yi} be the optimal programs of the local problems with i = 1, A:', and {xi} be the disaggregated solution components with i = l,fc. Form a new vector of aggregation weights ( {1 - 0)xi + 0xif at = { ( Xi,
i=TjF, (29.17) i = k' + 1, k .
Aggregation is feasible with 0 € [0,1], since xt > 0, x, > 0 for all i. The local constraints, which must be satisfied with the variables of the new aggregative problem, reduce to the following inequalities: Bi[(l-0)si+0£i]Xi
+ Diyi = bt,
Bix1Xz + Diyi - fi,
i=
Tj?,
i = k' + l,fc.
Hence it follows that the set Xi = 1,
i = \,k , i = k' + l,fc,
satisfies all the block constraints of the new aggregative problem {yi{0) > 0, since y~i > 0,yt > 0). Its remaining constraints reduce to the condition AQXQ = bo + pcrb, where
bo = b0 - 2 J AiXi =b0 - ^ i=i
i=i
k'
6b = } !=1
Aj(xi -
Xi).
ai Xi,
Decomposition
and
207
Aggregation
Let xo(p) be the solution of the parametric programming problem mincoa;0 ,
(29.18)
A0x0 = b0 + fiSb, x0 > 0. Then the set {x0(p),Xi,yl(p)}
(29.19)
is feasible in the new aggregative problem at least for those f3 for which solution (29.18) exists. However, with (3 — 0, this is the connected problem, therefore, by the condition of (29.1) there exists f3 > 0 such that (29.18) has a solution at least in the segment 0 < f3 < /3. The problem resides in estimating the solution of the functional of the new aggregative problem on its solution (29.19) with /? ^ 0. Assume that XQ > 0, i.e. the estimates UQ apply to a nondegenerate basic solution of the connected problem. In this case there exists 0' > 0, /?' € [0,/3], with which the basis of the vector XQ remains the basis of the optimal solution of (29.18) and with /? = /?', as well. Let x0((3') be the optimal solution of (29.18) with (3 = f3' being in the vector basis x0- Then we get c0xo((3') = u0b0 + /3'(u0,6b). And since u0 is the optimal solution of the problem dual to the connected, then c0x0(/3') =cox0 + (3'(u0,6b). The functional of the new aggregative problem on (29.19) is equal
$ = c0x0{p) + Y^ ci + ^2 d*yi(P) t=l
i=l k'
k'
= $ + P u06b + J3(Cj, xr - Xi) + ^{di, i=l
-& + yi)
i=\
k'
= (p + l3^{{uoAi
-Ci,Xi -Xi) + {di,yi - &)}
i=l
=
(29.19)
208
Systems
Optimization
Methodology
And since @ > 0 and z' < z', then, denoting the functional optimal value of the new aggregative problem by if', we get if* < if' < if <
Step 2: By the optimal solution of (29.3), setting 60 = b° - J2 ^ A i where 3^ is defined by formula (29.4), we construct the connected problem min J40£O =
COXQ , b0 ,
Xo >
0
and the dual to it max bou, U'AQ
> c0 .
(29.20)
Let u be the optimal solution of (29.20). Step 3: Define the vectors <Ti = UiAi — Ci and formulate local problems for each i = 1, k: max(cria;l + d;7/i), BiXi + Diyi = bi, »» > 0,
(29.21)
yi > 0.
Let (xi,yi), (i = 1, k) be the optimal solutions of problems (29.21).
Decomposition
and
Aggregation
209
Step 4: Define the values
i=l
i=l
where I, = <7i£i + dij/i, Z{ = OiX{ + dijji. Step 5: Test the optimality criterion. If z = z with a specified accuracy, then go to step 7. Step 6: Construct the set k' = {i : z, < ziti = Tj:} and define a ( i + 1 ) - a new vector of aggregation weights with the formula \ {1 - 0)Xi + 0Xi,
i = l,k>,
at = Let 1 = 1 + 1. Go to step 1. Step 7: Define the optimal solution of problem (29.1) {x*0,x*,y*},
i = l, A;,
where x* = ctiXi, i = 1, k. Terminate. §30. D E C O M P O S I T I O N A N D A G G R E G A T I O N B A S E D ON P E R T U R B A T I O N S M E T H O D The perturbations method is one of the major instruments of applied mathematics. Its basic idea lies in dividing a structure. It is this method that enables one to overcome a barrier existing between simple idealized models and complex real systems. Note that application of the perturba tions method calls for deep penetration into the essence of the problem in hand to isolate in it the main and secondary factors. §30.1. Consider the mathematical programming problem max
f0(x),
9o(x) < 0 ,
(30.1)
210
Systems
Optimization
Methodology
where x is the n-dimensional vector, go(x) the m-dimensional vector func tion. Concurrent with (30.1), consider the family of problems max(/o(:r) + e/i(:r)),
(30.2)
g0(x) +egi{x) < 0,
where e is a nonnegative scalar parameter. Problem (30.1) is identified as generative with respect to the family of problems (30.2) which will be termed the perturbated problem. The successful solution of the problem by the perturbations method is basically determined by two circumstances: firstly, the simplicity of properties of a generative problem, secondly, the proximity of the functions determining the original and the generative problem. Consider some types of the problems admitting immediate sim plification. (1)
max{F(zi(x1),...,zk(xk))/x'
6 Xs,
s= M} ,
(30.3)
where T{z\,... ,zk) is a monotone increasing function, x" are the ns vec tors, Xs are the subsets of the ns-dimensional Euclidean spaces. The solu tion of the problem reduces to the independent solution of the problem of the form max{zs(xs)/xs e Xs}, s = Ijfc. (30.4) If the generative problem takes form (30.3), it will be called the immediately decomposable. (2)
max{T{z1(x1),...,zk{xk))\g{z1{x1),...,zk{xk))<0}.
(30.5)
The solution of (30.5) is equivalent to the solution of the optimization problem in the aggregative variables Zi, (i = l,k). max{T(z1,...,zk)\g(z1,...,zk)
< 0} ,
(30.6)
here the solutions of problems (30.5) in the initial variables are the roots of the equations zs(xs) = z*s, s = TJc, (30.7)
Decomposition
and
Aggregation
211
where {z*} is the solution of (30.6). We will call the problem of form (30.5) the immediately aggregatable. (3)
max{F(z1(xl),...,zh(xk))\g(zl{xl),...,zk{zk))<0,
xseXs,
s-T7k}.
(30.8)
The set of its solutions constitutes the roots of equation (30.7). Here {z*} is the solution of the problem max{.F(zi,..., zk)\g(zi,...,
zk) <0, zs
where zs — mm{zs{xs)\x" G Xs}, zs = max. {zs(xs)\xa e Xs}. Here, as in (30.5), the monotonicity of T is not presupposed. It is evident that (30.3), (30.5) constitute generalization. The generative problem of such a structure will be referred to as that admitting immediate aggregation and decomposition. Suppose that we aim at constructing a solution of the perturbated prob lem (30.2) with a particular concrete value of the parameter e = £o- Assume that the immediate solution of the problem presents difficulties while the so lution of the generative problem proves to be comparatively simple. Then it is wise to substitute the solution of the original perturbated problem by the sequential procedure where the first stage constitutes construction of the generative problem, and the second is construction of a correction to it. §30.2. The study of connection between the solutions of the perturbative and the generative problem starts with considering the simplest situation where only the objective functional is subject to perturbation, i.e. we con sider the following problems: max(/0(i)+e/i(i)), x€ max
X; f0(x),
x € X.
(30.9)
(30.10)
Denote by X*,X*{E) the optimal solution sets of (30.9) (30.10), and by F*, F*{s) the functional optimal values of these problems. The apparently
212
Systems
Optimization
Methodology
evident statement about asymptomatic (with e — ► 0) nearness of solutions to (30.9), (30.10) can be refuted by the example. In the problem min{a;+£a;2 \x < - 1 } the solution x*(e) = m i n { - l / 2 e , - 1 } exists with any e > 0. And in the generative problem there is no solution. inf{z|x < —1} = —oo. We will set forth the theorem in which the conditions excluding the possi bility of similar inconsistency between the solutions of (30.9), (30.10) are isolated. T h e o r e m 30.1: Let the function fo(x) be continuous on X and the so lution set of problem (30.10) be nonempty. With any continuous function fi(x) satisfying the constraint sup fi(x) < oo for the solution set of problem (30.9) to be nonempty and bounded (with all sufficiently small e), it is necessary and sufficient that one could not isolate the point sequences {x{} such that Xi € X, \\xi\\ ► oo, lim fo(xi) = T* i—too
i — oo
The proof of the theorem as well as that of the following theorems are given in Ref. 30. T h e o r e m 30.2: Let the functions fo(x), f\{x) be continuous on the set X, X*(e) be bounded for all sufficiently small e, i.e. there exists e > 0 such that with 0 < e < E X*(e)cX, where X is the bounded, closed set in Rn Let the solution x* uniquely generate problems. Then for any solution x*(e) € X*(e) of the perturbative problem there exists the limit lim x*(e) = x*. £—0
As it follows from theorem (30.2), in the solution uniqueness case of the generative problem the range of searching the solutions of the perturbative problem can be narrowed to a certain neighborhood of the point x*. In this case, after solving the generative problem, the perturbations method reduces, first, to establishing the dimensions of the above neighborhood and obtaining by this technique the accuracy estimates for x* regarded as an approximate solution to (30.9), and, second, if it is possible, to constructing
Decomposition
and
Aggregation
213
corrections to this solution. The simplest estimates can be obtained on a basis of the following relations /o(**(e))+e/i(i*(0)>/o(**)+e/i(**), fo(x*(e))
< e{max{Mx)\x
(30.11) (30.12)
€ X) - fx{x*)} .
(30.13)
Relation (30.13) according to the function properties fo(x) and the set X may serve as a source of estimates for the value \\x* - x*(e)\\. We abandon now the assumption of uniqueness of the generative problem solution and consider the problem max fi(x), (30.14) x £ X*, the functional optimal value of which is Fx*. We will call (30.14) the auxi liary problem. Denote Argmax{/i(x)|a; £ X} = {y € XJx{y)
= max/i(x)}
Theorem 30.3: Suppose that all the conditions of theorem (30.2) have been fulfilled except for the assumption of the solution uniqueness of the generative problem. Then (1) the set X** of limit points of the elements X*(e) obtained with e —> 0, is contained in the solution set of the auxiliary problem X** C A r g m a x { / 1 ( : r ) | x € X * } ; (2) the objective functional optimal value T*(e) of the perturbative prob lem can be presented in the form
(3) on any solution of the auxiliary problem the objective functional value of the perturbative problem differs from the optimal on O(e). The consequence of theorem (30.3) represents the following two statements. Theorem 30.4: Suppose that the conditions of theorem (30.3) have been fulfilled and, moreover, the solution x** of (30.14) is unique. Then for any
214
Systems
Optimization
Methodology
solution of x*(e) (30.9) there exists the limit lim x*(e) = x** Theorem 30.5: Suppose that the conditions of theorem (30.3) have been fulfilled. Further assume that there exists i\ > 0 such that with 0 < e < i\ the solution x*(e) of the perturbative problem is unique. Then x*(e) is a continuous function on 0 < e < £i and there exists the limit lim x*(e) = x** e Arg max{/i(x)\x € X*} . §30.3. Consider a general formulation when not only the objective func tional is subject to perturbations but so are the constraints of the gene rative problem, i.e. consider problem (30.2). Assume that the generative problem (30.1) satisfies the regularity condition, i.e. at each point of the solution the constraint gradients fulfilled on the set X* as strict equali ties are linearly independent. It is known that, when fulfilling regularity conditions, the Kuhn-Tucker unique vector corresponds to each solution of the mathematical programming problem. Now, the solution of (30.1) corresponds to a unique nonnegative vector of the Lagrange multipliers A(x) = {\i(x),..., Xm(x)}. The auxiliary problem here takes the form max(/ 1 (x) - A'(x)g1(x)), (30 15) x£X*. We will set forth several theorems, 30 the statements of which form the basis for applying the technique of the perturbations theory. Theorem 30.6: Assume that the set X of the generative problem feasible solutions is closed and there exists e > 0 such that with 0 < e < e X*{e)cX,
X*CX,
where X is the bounded closed subset in Rn. Suppose that the functions fo{x), fi{x), go{x), g\{x) are continuous and have continuous partial deriva tives of the first order in a certain neighborhood of X and let the generative problem (30.1) satisfy the regularity condition. Then (1) the set X** of limit points of the elements of X*(e) obtained with e —+ 0 is contained in the solution set of the auxiliary problem, i.e. X** C Arg m a x i / U x ) - A'(x)9l(x)\x
€ X*};
Decomposition
and Aggregation
215
(2) the optimal value of T*{e) of the perturbative problem objective func tional can be presented in the form T*(E)
= T*
+ET?+0(E),
where T{ is the optimal value of the functional in (30.1) (3) any solution x* of the auxiliary problem can be set in agreement with the vector y(e) such that x{e) = x* + ey(e) 6 X(E), \\y(e)\\ < M = const., and on the solution X(E) the objective functional value of the pertur bative problem differs from the optimal on O(E). The consequences of theorem 30.6 constitute the following statements. Theorem 30.7: Suppose that the conditions of theorem (30.6) have been fulfilled and there exists E\ > 0 such that with 0 < e < E\ the solution x*(e) of problem (30.2) is unique. Then E% > 0 exists such that with 0 < E < £2, x*(e) is the continuous function E, here limx*(e) =x* G Arg max{/i(x) - A'{x)g1(x)\x
e—>0
€ X*}
Theorem 30.8: Suppose that the conditions of theorem (30.6) have been fulfilled and, moreover, the solution x' of the auxiliary problem is unique. Then for any solution X*(E) of the perturbative problem there exists the limit lim:r*(e) = x* = Arg max {.ft (1) - A'(x)gi(x)\x
6 X*}
£—►0
Thus, to obtain zero approximation of the perturbative problem in condi tions when the generative problem admits not one, but a set of solutions we need to isolate from this set the program which is the solution of the auxiliary problem. Consider the possibility of searching the perturbative problem solution as the series X*(E)
= X* + EX*X + ■ ■ ■ ,
(30.16) A*(£) = A*+£A! + --- ,
216
Systems
Optimization
Methodology
where x* is the solution of the auxiliary problem, A* = A(a;*) and calcu lating x*, A* will be described below. T h e o r e m 30.9: Suppose that the conditions of theorems (7), (8) have been fulfilled and, moreover, (a) the functions /o(x), go(x), fi(x), gi(x) are (m + l)-time differentiate in the neighborhood of the point x*; (b) the generative problem satisfies at the point x* complementary nonrigidity conditions; (c) it is nonzero, define det { W £ ( i * , A*), where C(x, A) = fo(x)—A'go(x), {W£(xr, A*)} is the matrix of the function C(x,k) second deriva tives by the x vector components and the positive components of the vector A. Under these conditions expansion (30.16) exists. Theorem 30.10: Suppose that the conditions of theorem (30.9) have been fulfilled, but instead of the strict complementary nonrigidity conditions and the nondegeneracy condition VV£(i*,A*) we presuppose sign definiteness of the matrix VxVxC(x", A*). Then expansion (30.16) exists. We will describe the procedures of obtaining x*, A*. In conditions of theorem (30.9), x*, A* are calculated from the formula
(g)={w^.A-»^ : * : : : : ^ : *: : « ; : : : ;^). ( «u7) where y^,^; will be described below. The regularity conditions in the generative problem imply that the regu larity conditions are also fulfilled with a sufficiently small e in the perturbative problem. Therefore, with these e there exists the Kuhn-Tucker vector A*(e) satisfying the equations: Vf0(x*(e))+eVh(x*(e))
= £ \t (e)(V gh(x* (e)) + eVgh (x*(e)), ie« (30.18) glo(x*(e))+egh(x*(e)) = 0, lew,
where w is the set of constraint indices fulfilled on x* as strict equalities. This system can be regarded as an implicit specification of the function
Decomposition
and Aggregation
217
x*(s), A*(e), I € w in the neighborhood of the point (x*, A*, / 6 w, e = 0) and, considering that det(VV£(:r*, A*)) ^ 0, on the strength of the wellknown theorem of the implicitly specified functions we come to expansions (30.16). And the relations of (30.17) as well as the explicit expression for V?i, ipi can be obtained by substituting (30.16) in (30.18) and expanding appropriate functions in a series according to degrees e. Thus, in the case when the set of constraint indices of the perturbative problem fulfilled on the solution (with a sufficiently small e) as strict equalities is known, the procedure of searching expansion coefficients is described. But, in general, when analyzing the generative problem the set of the identified cannot be available. In this situation we substitute the formal expansion of (30.16) in the functional of the perturbative problem to obtain fo(x*)+e(Vf0(x*))'xi / + e I -xlVV/0(x*)x1
\
2
T
+(Vf1(x*)) x1
+(V/0(I*))'I2
(30.19) )+••••
Evidently, x\ must be the solution of the coefficient maximization problem with e max{(V/ 0 (x*)) T x 1 |(V 5 / o (i*))'x 1 < -gh(x*),
I € ft(x*)} .
(30.20)
At the same time, the solution of (30.20) is constituted by the vector x\ satisfying the relations (Vgi0{x*))'x1
= -gh(x*),
lew, (30.21)
(V
' e ft(z*) \ w ,
where w is the set of the positive component indices of the Kuhn-Tucker vector, Q(x*) is the index set of the constraints fulfilled on x' as strict equalities. Denote by X^ the vector set satisfying (30.21) and fii(xi) the set of constraint indices in (30.21) fulfilled on xi € X{ as strict equalities. Then x*, x\ represent the solution of the problem 1 max{ ^xiVVfoix^x. 2
+ iVhix*))1
V9l<j{x*)x2 < -^x\VWglo{x*)xl
x1+(Vf0(x*)y
x2\Xl
e XI
i
(30.22) T
- (Vgh(x*)) xu
I 6 U^Xy) I ,
218
Systems Optimization
Methodology
providing the coefficient maximum with e2 in the perturbative problem functional writing. It is evident that
max\{Vf0(x*))'x2\Vglo(x*)x2
< -2*1 VV f t o (i*)n - (Vgh(x*)y xi, I G f2i(ii) *\T = (A*)
-^iWjJl'JU-^ll'))'!!
(30.23)
And, therefore, problem (30.22) can be written as max <
-xfVXVxC(x*h*)xi
(30.24) V
I6«i
/
If the problem solution is unique, then it coincides with x\. Furthermore, the Kuhn-Tucker vector K\ of problem (30.24) coincides with the coefficient with e in (30.16). To obtain x2, A2, we need, by utilizing transformation similarly to (30.23), to reduce the coefficient maximization problem with x2 to the problem of maximizing the functional which depends only on x\. §30.4. Consider application of the perturbations method to LP problems. The nonuniqueness of the generative problem solution introduces changes into the procedure of constructing approximate solutions, since the problem arises with respect to isolating that of the generative solutions to which the perturbative tends as the parameter e decreases. The nonunique ness is particularly characteristic for linear problems. Thus, we have the generative problem max coXo , A0x = bo,
x >0
(30.25)
Decomposition
and Aggregation
219
and the perturbative problem max(c 0 + ec\)x, (A0 +eA1)x
= b0 +ebi ,
(30.26)
x > 0, where A0,Ai are the (m x n) matrices, c0,ci,x are the n vectors, b0, 6j are the m vectors. We assume that the rows of the matrix A0 are linearly in dependent. Then the regularity condition for problem (30.25) is equivalent to the assumption that the basic components of each optimal solution of (30.25) are positive. In this case, the dual problem for (30.25) has a unique solution A*. Consider the auxiliary problem max{(cix) - (A*, Axx)\ ,
(30.27)
x £ X* , where X* is the solution set of (30.25). We will rearrange for our case theorem 30.6.
T h e o r e m 30.11: Let the set X*(e) of the perturbative problem solutions be constrained and the optimal solutions of the generative problem be nondegenerate. Then: (1) the set X** of limit points of the perturbative problem elements X*(e) is contained in the solution set of the auxiliary problem, i.e. X** C A r g m a x { c i a ; - (A*,Aix)\x
G X*};
(2) for the optimal value P*(e) of the perturbative problem objective func tional the following expansion holds
f*(e) =?*+
e(T* + A*r&0 + 0(e2);
(3) if x** is an arbitrary basic solution of the auxiliary problem and A is its associated basic submatrix of the matrix A, provided the columns corresponding to the nonbasic variables x** are replaced by nonzero columns, then with all sufficiently small e the basic program x(e) = (xt(e),0), Xb(e) = (Aob + eAib)~l (b0 + ebi) is feasible for the
220
Systems
Optimization
Methodology
perturbative problem and the functional value on it differs from the optimal on 0(e2). Formulate the perturbations method algorithm for the LP problem. Step 1: Define the solution set X* of the generative problem (30.25). Step 2: Obtain the solutions of the auxiliary problem (30.27) and fix the indices corresponding to basic variables. Step 3: As the solution "suspicious" for optimality, in the originally (per turbative) problem (30.26), we take for the basic program the basic variable indices which coincide with those for the auxiliary problem. Step 4'- The solution is tested for admissibility to problem (30.26). As it follows from theorem (30.11), if e is sufficiently small, then it is feasible. Otherwise, the method of perturbations cannot be applied. Step 5: Calculate characteristic differences and test the program obtained for optimality. Step 6: If the program is not optimal, i.e. either the auxiliary problem solution is nonunique or insufficiently small, then the basic pro gram obtained can be employed for further calculations that can be carried out, say, with the use of the simplex method. §30.5. By using the algorithm described, we will provide several examples and demonstrate beneficial properties of separate problem types. (A) Consider the following LP problem
I J=I j 6 n , (A + eB)x = b,
J
(30.28)
x>0,
where A is the matrix of a block diagonal structure and Aj (j — 1,£) are its blocks, the vector b consists of m3-dimensional vectors bj, the vector x = (xi,... ,xc)- Note that the generative problem for (30.28) is of the immediately decomposable type (30.3).
Decomposition
and
Aggregation
221
The algorithm of the perturbations method for problem (30.28) is im plemented as follows: (1) Solve £ subproblems of the type (30.4) m a x
l ^2 CjXj\AiXi = bu xt >0>, I J6Q| J
l = TjC.
(30.29)
If the solution of these problems is unique, then we get the solution of (30.28) accurate to 0(e). To refine the solution obtained, we fix the optimal basic sets of the indices <7/ (I = 1,£) of each subproblem and carry out transition to point (3). If in some problem of (30.29) the solution is nonunique, which appears in even one of the nonbasic characteristic differences being equal to zero, then the optimal solution sets X* of these problems are isolated and the vectors A;, / = 1,£ of the optimal dual variables, which are under the regularity assumption are unique and are fixed. (2) Solve the auxiliary problem which falls into £ independent subproblems min I J2(Aa,Baixt)\xi
€ X,* 1,
l = TjC,
(30.30)
where Bsi is the submatrix B in which the rows correspond to the components 6 s , 5 = 1,M of the vector b, and the columns correspond to xi. Moreover, A s , 5 = 1,M are here the components of the vector c A = ( A i , . . . , Ar) and M — ]T nij. Assuming that all the problems j=i
of form (30.30) have unique solutions and denoting them by x\, we can take as the approximate (accurate to O(e)) solution of (30.28) the vector x* = {x*i, l = Tj£) To refine the solution, the index set of the basic optimal bases cri is fixed in problems (30.30). (3) We set in the system (yl + eB)x = b all nonbasic variables being equal to zero, i.e. c xfc = 0 , k$[Jai (30.31) 1=1
Systems Optimization
222
Methodology
whereupon it takes the form (A + eB)x = b,
(30.32)
where x is the vector composed of variables with basic indices. If sys tem (30.32) with the specified e has the non-negative solution, then it, together with (30.31), forms a feasible solution of the original problem. To test the original problem solution for optimality, we solve the system of equations for the estimates (A + eB)A = c, where c= I ck,k 6 (Jcr; > ■ Then we test fulfilling the inequalities (.4 + eB)A > c, which enables us to see whether the fixed e belongs to the segment indicated in the conditions of (30.11). (B) Consider one more problem of the form c m a x y^CjXi
,
;=i
A\x\ + eBix = bi + ui,
c \] ui ^ £^o >
(30.33)
xi 6 Xi, %n > 0, I = 1, £ , where x = (x\,..., xc), u3, j = 1, C are the vectors of the unknowns Ai, Bt, bi (I = 1, £), 6o are the matrices and vectors of the correlated dimension alities, and Xi are polyhedrons. One cannot apply the algorithm of the perturbations method immediately to (30.33) since, due to the structure of the constraint first group, the generative (with e — 0) problem obviously is irregular. By changing the variables u; = evi, we get the problem £
max
2_\cixi. z=i £
Aixi + eBtx = bi + evi,
; P vi < b0 , i=i
xi € Xt,vi > 0, / = 1,£.
(30.34)
Decomposition
and
Aggregation
223
The generative problem for (30.34) has the form £
2_,ClXl
max
1=1
Am=bi,
xiZXi,
(30.35)
£
^2vi
= l,C-
Problem (30.35) has no unique solution. Indeed, even though all the problems max{axi\Aixi =bhxt € X[} (30.36) have the unique solutions x\, the variables vi in (30.35) remain indeter minate and, to secure their approximate value v*, we need to solve the following problem: £
max
A vi Y2 ' ' i=i
£
y^vi
(30.37)
< bp, vi > 0,
i=i
where A; are the solutions of the problems dual with respect to (30.36). And problem (30.37) has the simple solution (bsQ, v* = < s
|0,
l=Ps,
l^Ps,
ps = a r g max A?,
where Xf are the components of Aj, 6Q a r e t n e components of bo- Thus, in the initial approximation, the resource S is transferred to the subsystem which employs it on the program x* most efficiently. Next, following the method of perturbations, one solves the system Aixi + eBix = bt+ ev*,
(30.38)
where xi is the vector involving the variables xi with the index from the optimal basic set of (30.36), and Ai, B\ are the appropriate submatrices of the matrices Ai, Bu Evidently, as before, if system (30.38) admits a unique
224
Systems
Optimization
Methodology
non-negative solution with the specified concrete e, then it is also a feasible solution of the original problem. (C) And, finally, we will consider the problem close to the directly aggrega tive (30.5): C
max ^2^2(d+
e&i)xj ,
1 = 1 j€fii
(30.39)
c
J2 X! (Ai+eBi)xi
°.
1 = 1 jGfii
f3Jt
where Xj, cj, are scalar vectors, At, Bf, b are m vectors. Introduce the notations 2; = 5Z xj a n < i rewrite problem (30.39) as
max I X cizi + £ X V 1= 1
c c ^2 Aizi + £ ^ ^ (=1
Y^ $xi
i = l jGfii
) ' /
(30.40)
$Xj < b; x > 0; z > 0.
Z = l j€Cli
Let (z*) be a unique solution of the aggregative problem c max 2_\ciZi, 1=1
c Y,Aizi
(30.41) z
i>°-
i=i
Then the optimal solution set X* of the generative problem is determined by the conditions 2 J Xj = z\, Xj > 0, j € fli, I G a, Xj■ = 0,
j e Hi,
IGCT ,
where a is the variable index set entering into the optimal basis of prob lem (30.41). Let A be the solution of the problem dual with respect to (30.41). Based on theorem (30.11), we can state that the solution sought
Decomposition
and Aggregation
225
coincides, accurate to 0(e 2 ) (by the functional), with the solution of the auxiliary problem max
\ 5Z Y, W? ~ A'Bfta^l* eX*\,
(30.42)
which, due to separability, falls into |cr| element problems (I = TTcr): max] ^ ( ^ - A ' B / ^ l x e X * } , J ] *i = Zi,Xj
>0.
Let us formulate the perturbations method algorithm as applied to prob lem (30.39) Step 1: Obtain the solutions z", A of problem (30.41) and its dual problem. Step 2: Let ji = Arg max (/?/ - A'Bj), I G
Xi =
z*,
{o,'
;=ii.
' e
i
if";
T^ iii
i e H/,
lea;
Step 4: Since the optimal basic set of (30.42) is formed from the indices ji, I & a, then the nonzero components of the locally exact solution can be obtained from solving the system
Y^{Ai+eBls)xjs
= b,
(30.43)
(Go-
where Ai, Bi are the submatrices of the corresponding matrices from which all the rows with the indices of inequalities in (30.41) not transforming into equalities on z* are removed. If ji are determined ambiguously, then the difference of the functional from the optimal on 0(e2) can be ensured on the solu tion of (30.43).
Systems Optimization
226
Methodology
§31. DECOMPOSITION M E T H O D BASED ON AGGREGATION OF VARIABLES FROM DIFFERENT BLOCKS We will describe the method 40 applicable to a wide class of extremal problems, in which the variables entering into different blocks are aggre gated, as contrasted, say, to the method described in §29 of the present chapter where macrovariables correspond to each block. §31.1. Let us describe the sector planning model taken as an example to consider the aggregation method. Let a sector (industry) be specified by the following combination of sets: k ■=■ [1 : K] is the set of the sector plant numbers, M = [1 : M] is the number of nomenclature of the items produced at the sector plants, kj is the set of numbers of the equipment groups available at the plant j , Ij C / is the number set of the complete items produced at the plant j . Assume that the technological chain of producing a particular article at a particular plant is rigidly fixed. Then the production capacities of each plant within the model of employing the noninterchangeable groups of equipment are described by the following inequalities Yl T]kx) < $ ) ,
k € kj,
x)>0,
i £ Ij ,
(31.1)
iei,
where rrj is the yearly output of the articles i at the plant j , $* is the yearly fund of operation time of the equipment group k at the plant j , T 1 denote the expenditures of operation time of the equipment group k at the plant j for producing one article i. From inequalities (31.1) we see that the values xj are determined for each plant j only for i € Ij. Furthermore it is convenient to define them for all z £ f setting ar*. for i £ I \ Ij. The connection between the output of complete articles and the sector final output is described by the relations k
M
j= \
771=1
53(*j + toj)= ^ r - r m ( i + 5m),
(3i.2)
where Ym is the yearly final output of the form m, wz- is the inventory of the articles i at the plant j at the beginning of the planned period, l'm
Decomposition
and Aggregation
227
is the number of the form i entering into the output of the form m (the complating coefficient), Sm is the standard of the output inventory of the form m shifting to the next planned period (in unit fractions). Assume that the set of the final output numbers M is divided into two nonintersecting subsets: M = MiUM2,
M = [1:M1],
M2 = [Mx + 1 : M].
The products from the set Mi are the most important for the sector from the standpoint of the national economic demand, the output of these products is compulsory and their volumes are either fixed or specified by constraints from below Ym>Ym, me [I: Mi}. (31.3) Assume that there exists the vector { ? M l + 1 , . . . ,YM} such that the vector {Y\ ..., YM\ YMl+1,..., YM} satisfies (31.1), (31.2) for some x). As far as the products from the set M2 are concerned the decision making with respect to their volumes is within the competence of the planning body of the sector (industry) itself. This part of the nomenclature is called free. Let the planning body set the free nomenclature vector {YMl+\...,YM}
(31.4)
Two situations are possible here. (1) Vector (31.4) is unfeasible, i.e. there are no values such that conditions (31.1), (31.2) are fulfilled. (2) Vector (31.4) is feasible. Here, generally speaking, an infinite number of techniques is available at the plants and a particular part of the sector capacities may not be involved. In both cases, the planning body of the sector encounters the necessity to reconsider vector (31.4). If vector (31.4) is regarded as the notion of the wanted structure of the free nomenclature of the sector final output, then the final plan must be sought for as YMl+1 =eYMl+1,...,YM
=SYM
.
(31.5)
The components of vector (31.4) are called the assortment relations, and the magnitude 0 is identified as the number of assortment of M sets. Thus, if
228
Systems Optimization Methodology
vector (31.4) is assigned, then we need to solve the following optimization problem: 0 has to be maximized under conditions (31.1) - (31.2). We rewrite (31.2) as Jb
Mi
M
53(*5+«j)= ^ / i m y m ( l + <5m) Y. llmYm(l + 6m),i = YJ j=l
m=l
m=M\ + \
or, utilizing (31.5), we get fc j=i k
where W1 = Y2 w1, is the summary pool of the article i in the sector at the Mi
beginning of the planned period, H{ = £ llmYm{\
+ 6m) is the summary
771=1
volume of the article i needed to produce the final product of the compulsory nomenclature, a' =
llrnYm(l
Yl
+ 6m) is the ith component of the
m=Mi+l
assortment relation of the complete article output. Denoting V1 = Wx — H*, i = Tj; bf = Tj fe |#*, j = ljfc; k e k,; i € I} we will finally write the optimization problem as max 0 , (31.6)
^2bfx)
(31.7)
ieij
x) > 0, i £ Ij, x) = 0, i 6 I\If, j € [1 :
fc],
(31.8)
A:
aJ0-£>} =V\ie[l,J],
(31.9)
0>O.
(31.10)
Problem (31.6) - (31.10) is an LP problem the dimensions of which are enormous in the real problems of sectoral planning. At the same time, the problem has a block structure: constraints (31.7) - (31.8) refer to se parate plants and are block constraints, constraints (31.9) are sectoral or connecting. §31.2. Let us describe the concept of the decomposition method for prob lem (31.6) - (31.10).
Decomposition
and Aggregation
229
We introduce new variables - complete outputs as per all plants: k
JC* = J 2 4
i=ij.
(3i.ii)
The values Xx are identified as aggregative variables. We also introduce specific outputs of the article i at plant j a)=x)/X\
je[l,k],
ie[l:I]
and call the values a1- the aggregation weights. Evidently, the conditions k
5 > } = 1, U a*. > 0,
ie Ij,
i6[l:/], (31-12)
a) = 0,
iel\ljy
je[l:k],
are fulfilled. Assume that the aggregation weights satisfying (31.12) are specified. Then, substituting in (31.6) - (31.8) the variables x) =■ a} Xi, j 6 [1 : k]\ i E [1 : I] and denoting S*fc = &}fco^, j g [1 : k], k 6 kj, i 6 Ij, we write problem (31.6) - (31.10) in the form (31.13)
max 0 ,
j€[l:i],
J2B;"X><1,
a'0 - X* = F 1 ,
Xl>0,
ie[l:J] >
0>O.
rC t
Kj ,
(31.14) (31.15) (31.16) (31.17)
We call problem (31.13) - (31.17) the problem in aggregative variables or the macroproblem.
230
Systems
Optimization
Methodology
The dual problem for (31.6) - (31.10) has the form
n r m ^ E E ^ + E^V, X>^*-/^>0, fce^j
,'€[l:i],
i£l3; (31.18)
/ i=i
#>0,
j € [1 : * 1 ,
feefcj-
And the dual problem for (31.13) - (31.17) is as follows
j = l k£kj
1=1
k
^ B " ' f
V>0,
ie[l:/],
(31.19)
j = l kek, I
£VA*>1,
i#>0,
j'G[l:Jfc],
kekj.
i=i
We denote the optimal solutions of the problems by the sign * above the corresponding letters. Let X1, i £ [I : I] be unique optimal solutions of problem (31.19). For each fixed index j we formulate the block (plant) problems max hj — 22 ^x) > Y,b)kx)<\,
(31.20)
fcefc,-,
i£l,
x} > 0,
ie/j,
x)=0,
i£l\I3
Decomposition
and
231
Aggregation
The dual problems for (31.20) with each j 6 [1 : k] are written as
min ft = £ tf , (31.21) €J > o, fc e Jfcj
•
The iterative process is constructed in the following way. With some fixed weights of a* satisfying (31.12), we solve the problem in aggregative varia bles (31.13) - (31.17). It will be shown below that the problem, under some assumptions, is solved analytically. Further, by using the dual estimates *. A1, i € [1 : / ] , the functionals of block problems (31.20) are formed and these problems are solved. Let x j , j 6 [1 : k], i € [1 : / ] be their optimal solutions. Introduce the variables x)=a)X\
je[l:fc],
i 6 [1:1],
*. where X%, i € [1 : / ] are the optimal solutions of problem (31.13) - (31.17). The values x1, are termed the disaggregative solutions. Now aggregation weights axe defined as the function a^pj) according to the relation
£[l} + P i (zj-lj)]
a)(Pj) =
3=1
,i€L
(31.22)
I 0, i 6 / \ Ij , where 0 < Pj < 1, j € [1 : k]. Evidently, conditions (31.12) for the weights aUpj) are met. If we consider problem (31.13) - (31.17) with weights (31.22), then the optimal value of the functional 0 in it represents a * function from the parameters Pj-&{pj)- The problem arises with respect to maximizing the function Q{pj) under conditions 0 < Pj < 1, j 6 [1 : k]. Let the maximum be achieved with some p3, j 6 [1 : k]. Then the weights for the next step of the iterative process are determined from formula (31.22)
232
Systems Optimization
with pj = pj. *
l
Methodology
Thus, the algorithm forms a sequence of disaggregative
■
*
solutions x j which together with 0 are admissible to the original problem (31.6) - (31.10). §31.3. Establish the optimality criterion of the solution {x*, 0 } for problem (31.6) - (31.10). Assume that with some weights of a} satisfying (31.12) we have obtained the optimal solution {X 1 ,0} of the problem in aggregative variables *. (31.13) - (31.17), and x* is the appropriate disaggregative solution. Let \] be a unique optimal solution of the dual problem (31.19). Denote hj = £
\% ,
(31.23)
XZ) .
(31.24)
i€lj
hj = £
The values hj, hj constitute the functional values of block problems (31.20) on the optimal and the disaggregative solution, respectively. Let h = ^2 hj, h = ^2 hj, The following theorem gives an optimality condition of the disaggregative solution. Theorem 31.1 4 0 : Suppose that for the optimal solutions {X*,Q}, Xi, i € [1 : / ] of direct and dual macroproblems and the optimal solutions £', j € [1 : k], i € Ij of block problems the following equalities are fulfilled hj = hj,
je[l:fc]. *
■
*
(31.25) *
■
■
*
Then the disaggregative solution {x'j,0}, where x1 = a)Xl; i € Ij is an optimal program of the original problem.
■
j £ [1 : fc],
§31.4. Since the central moment of the method described is represented by repeated solution of the problem in aggregative variables (31.13) - (31.17),
Decomposition
and Aggregation
233
then it is essential to ascertain the properties of the problem solution. We express X{ in terms of 0 according to (31.15): X* =a«0-V\
(31.26)
ie[l:J]. ie[l:J].
By substituting (31.26) in (31.14), obtain
eY,B?ai-J2BTVi
ie[l:*],
kekj.
ti=l =l
From these inequalities we derive the expression for an optimal value of the variable 0 :
0 = mm ll + ^B^VA kek
\
i=\
/( f^B^aA /1 \ I = i
..
(31.27)
/
*. * And the optimal values X\ i € [1 : I] are determined in terms of 0 from formula (31.26). Consider now the dual problem (31.19). Assume that the output of each article is strictly positive: X£t>0,
i€[l:I]. ie[l:J].
(31.28)
According to the second theorem of duality, if the adjoint problems are solvable, then in each pair of their dual conditions one condition is free and another is fixed. Hence and due to (31.28) the first group of relations of (31.19) is fulfilled on the optimal solution as equalities, i.e.
^EE^
(31.29)
(31.29)
j=i fcefcy
Substituting (31.29) in the functional and the second group of constraints of problem (31.19), we come to the problem
*#-tE*(^t^), i=i
fee*,
\
i=l
(31.30,
(31.30)
/
(31.31) =l j= = ll k£kj fc€fcy \\ tt=l
/
234
Systems
Optimization
Methodology
Let the minimum in (31.27) be attained on a unique pair of indices (j*, k% i.e.
© = (i+J2 zy vi J / ( E BT A ■
<31-32)
In this case problem (31.30) - (31.31) has a unique solution.
$A£ = I / ( £ B £ * A
$=0,
(j,fc)^(j*,F).
(31.33)
*. And the values A1, according to (31.29) and (31.33) are equal
V = Bf.' I I £ - B £ V " j , i £ [1 : 7]
(31.34)
and determined unambiguously. This is a simple case. Let the minimum in (31.27) be attained on a particular set R of index pairs (j, k). Then (31.27) can be written as 0
= (1+ EB?V1) / ( E B > 1 ) -
0',k)£R-
For the values 77^ we have the following problem
E ^(i+I>ikH^min
(i.*)e
E ^ ( E ^ > ' ) = I. ,fc)efl
#=0,
\ »=i
/
( 31 - 35 )
(31.36)
{j,k)ZR.
The problem has a nonunique solution, and the question arises with respect to determining the values A; which form the functional of block problems.
Decomposition
and Aggregation
235
This case is referred to as degenerate. Thus, consider problem (31.35) (31.36) when degenerating. A set of its solutions belongs to a polyhedron with nodes ^°
=
1
/ ( E
S
V
; °
(Afc°)ei?.
) >
(31.37)
Enumerate successively all the elements of the set R and denote the resul tant set of numbers by s = [1 : a]. Substitute (31.37) in (31.39):
K=Bf
(£S>M,
U,k)ER,
sES.
Thus, the solutions of the dual macroproblem can be presented as s
^' = X)7.A'„
(31.38)
s=l
where the values j s belong to the set G= \-y = (lu...,ls),J2ls
= hls
>0, 3 G [ l : s ] l .
3=1
Denote by M J with each j € [1 : k] the sets of feasible solutions of block problems (31.20). Consider the following problem: max min(/i - h) = max min 2 J 2 J ^ J i ^ A ^ x J - I } ) . x-€MJ j€[l:k]
l€G
*)eM' 1eG je[l:*]
j=1
ieIj
(31.39)
s=1
If M J , / € [1 : k] are the bounded sets, then problem (31.39) has the saddle point (x1-, 7 S ). Here for each j the values xl-,i E Ij are the optimal solutions of block problems (31.20) with the functionals in which *
■
*
—
*
3= 1
*
•
236
Systems
Optimization
Methodology
Denote in terms of h the value h for the saddle point x*. Thus we come to the following optimality criterion for the disaggregative solution x* in the degeneracy case. T h e o r e m 4 0 : The sufficient condition for optimality of the disaggregative solution {x*j, 0 } for the original problem (31.6) - (31.10) in the degeneracy case constitutes the equality h = h.
(31.40)
If problem (31.6) - (31.10) has a solution, and the program {x^, 0 } is _ * not optimal for the problem, then h > h. Note that the non-uniqueness of solutions of dual macroproblems (31.19) presents one of the difficulties encountered when implementing the iterative scheme described. Reference 40 emphasizes the fact that degeneration of problems (31.19) becomes sub stantial when approaching the optimal solution of the original problem, and in the numerical experiments conducted the optimal dual variables A1 were calculated proceeding from the fact that the minimum in (31.27) is attained on a unique pair of indices even when approaching the optimum of the original problem. §31.5. Let us formulate the algorithm of the decomposition method de scribed for solving problem (31.6) - (31.10). Step 1: Introduce some fixed aggregation weights a ' conditions
satisfying
the
k
5 > j = l,
iG[l:J],
3=1
Q}>0,
IEIJ,
aj=0,
i€l\Ij,
j € [1 : k].
Let the iteration number 1=1. Step 2: Solve the problem in aggregative variables (31.13) - (31.17). We know (§31.4) is the analytic solution of the problem
Decomposition
fcefc, J \
and Aggregation
1=1
i^a'0-V
1
// ,
\«=i
237
/
(3i.4i)
i€[l:J].
Step 3: Obtain the dual estimates of A\ i € [1 :1] to form the functionals of block problems (31.20) from formula (31.34)
h=
B;t'/\J2B;ra\ie[i:i],
where (j*, k*) is the pair of indices on which the minimum in (31.41) is achieved, and in the degenerate duality case of the macro*. problem, i.e. when A1 is nonunique, any one of the optimal solutions A1 is taken. Let £*■, j' £ [1 : k], i 6 [1 : / ] be the optimal solutions of problems (31.20). Step 4- Obtain the disaggregative solutions x)=a)X\
j €[!:*],
i 6 [1 : / ] .
Step 5: Test the optimality criterion of the solution {x1-, 0 } of the original problem. If * hj = hj, j e [ 1 : k], * where hj, hj are calculated from formulae (31.23), (31.24), then the solution is optimal. Go to step 8. Step 6: By using formula (31.22), we form the conditions of prob lem (31.13) - (31.17), the functional of which is afunction from the parameters pj, j G [1 : k], and solve the problem with additional constraints 0 < pj < 1, j € [1 : k]. Let pj, j 6 [1 : k] be the optimal solution. Step 7: From formula (31.22) we define new aggregation weights al}{p*). Set I = / + 1 and return to step 2. Step 8: Terminate.
238
Systems
Optimization
Methodology
In conclusion it should be noted that the above decomposition method based on aggregating variables from different blocks can be applied to a wide class of extremal problems, e.g. problems of linear and square programming with a block diagonal structure of part of constraints, the optimal control problems with mixed constraints, etc.
Chapter 8 A P P L I C A T I O N OF SOLUTION T E C H N I Q U E S FOR LARGE D I M E N S I O N PROBLEMS TO G R A I N FARMING OPTIMIZATION
§32. GRAIN CHAIN §32.1. The Russian government is facing, say, the problem of developing the long-range plan of an appropriate complex by structural products including the national economic indicators prescribed from above and satisfying the national economic needs with final products of this diversified complex with minimal expenses. It is impossible to construct a model embracing all the problems of agriculture and food processing industry with all the existing relations with other national economic industries. We can only develop a system of models, wherein each model represents the mathematical formulation of one or several closely interrelated problems abstracting from all other factors which actually affect the problem. The system of models is constructed so that, proceeding from the prediction and general economic limitations on investments and labor force, the optimal distribution of basic funds and the labor force according to individual structural products and the limitations on imports are determined and the national economic demand for these products is established. Parallel planning is then carried out by all structural products: meat, milk, sugar and starch, grain, etc. The obtained results are balanced with each other and with other national economic industries, the efficiency of such a plan is determined and these new results enter again into the system as 239
Systems
240
Optimization
Methodology
initial data. The entire process starts again. In other words, the process of planning is the iterative one. One of the main problems is to carry out the process of planning struc tural products by mathematical methods, i.e. construction of the so-called chains. In this case the grain chain is an example for constructing other chains. §32.2. The grain chain describes grain production and its subsequent con sumption in agriculture itself and in industry before the first processing stage. Grain production and consumption may be represented schemati cally as follows: Here fertilizer
r~
straw
r T
means the following: Not all the straw obtained from grain production as a by-product is needed for litter and industry. The remaining part is left cut into pieces in the field. After ploughing it turns into fertilizer. preliminary grain cleaning preservation grain drying
I
storage
These processing stages are determined by specific agricultural conditions. Upon storage the natural properties of grain are preserved by reducing vital activity of grain. Humidity and relatively high temperatures axe the essential factors affecting the grain's vital activity. If the water content in grain is more than 14.5 percent, its vital activity increases markedly and its quality simultaneously deteriorates. The cost of drying units is rather high, and they are used during harvesting. Because of this, from the viewpoint of the main fund utilization it may be beneficial to extend drying time by preserving grain. Three preservation techniques are considered here:
Application
of Solution
Techniques for Large Dimension
Problems to . . .
Grain Production and Consumption
APC - Agricultural Producers' Cooperative
241
242
Systems
Optimization
Methodology
(1) Grain is aerated in suitable rooms by cold air at a temperature of —1° to +1°. This is what is known as preservation by air supercooling (maximal storage life is 30 days). (2) Grain is aerated by external air. This is what is known as simple preservation (maximal storage life is 14 days). (3) Grain is in the open air. It is not treated. (Storage life is 7 days). §32.3. The national economic demand for grain will grow progressively at the cost of increase in the livestock output. The need of population for cereal food products remains approximately the same. Russia has to import large quantities of grain from abroad, since its own grain production cannot satisfy this need. Thus, the problem arises with respect to growing high-yield crops, increase of fertility at the expense of land improvement, increase in drying and preservation capacities, increase in storage volume, enlargement and improvement of the fleet of combines, tractors and lorries, etc. This aims at providing Russia with its own grain. We can construct the following grain chain: — Proceeding from forecast, we determine the need for fodder. For the last year of the long-range plan we establish a fodder balance which results in obtaining the need for grain fodder. — Based on the forecast, we determine the needs of the population for food products for the last year of the long-range plan. Proceeding from the knowledge of such needs, we have to construct the models optimizing satisfaction of needs by different criteria since a large number of food products are interchangeable. Starch may be taken as an example here. Though starch is produced from potato or grain, in a sense women are indifferent to the kind of starch used in the kitchen, irrespective of whether it is produced from potato or grain. Grain is the weakest point in the Russian national economy. We would like to determine whether a particular portion of starch produced from grain can be made from potato and what expenses are associated with the problem, since equipment will require readjustment. Implementation of such models will result in the knowledge of demand for cereal food products.
Application
of Solution
Techniques for Large Dimension
Problems to . . .
243
— The forecast touches upon requirements for growing new grades of grain. Here we deal with the requirement for yield capacity, quality of the grain itself, immunity against diseases and the like. — Field capacity depends on the type and quality of fertilizer introduced per 1 hectare and on the extent to which land is improved. It is also determined by the application of pesticides to protect plants and by many other factors, which cannot be estimated qualitatively. §32.4. The greater the yield capacity, the greater is the income obtained from 1 hectare. It can be presented by a concave function according to these indicators. Moreover, the amount of fertilizers, pesticides for plant protec tion and the sum of investments for land improvements are constrained by grain farming. Hence it is wise to construct production functions for each grade determining the dependence of yield capacity upon the three factors, i.e. construct fi{xi,yi,Zi)
- the ith grade yield capacity, where
Xi - the amount of fertilizers introduced per 1 ha, yl - the sum of investments per 1 ha employed for land improvement, Zj - the amount of toxic chemicals utilized, and we have to determine with their use the optimal yield capacity under given conditions. Let pi be a procurement price of 1 centner of the ith grade grain, q - the price of 1 centner of fertilizers, v - the price of 1 centner of pesticides, bi - the average quantity of fertilizers introduced per 1 ha of the ith grade grain, Ai - the area sown with the ith grade grain, A - the area sown with grain, d - the amount of investments for land improvement, which accounts, on the average, for 1 ha of the area sown with the ith grade grain, a, - the quantity of mineral fertilizers accounting, on the average, for 1 ha of the area sown with the ith grade grain. The yield capacity optimization problem then takes the form: n
^2{Pz
- qxi - yi - vzi) ->max
244 Systems Optimization
Methodology
Application
of Solution
Techniques for Large Dimension
Problems to . . .
245
thus
i>=y'
ji>=x, =1
i=\
J2zi = Z,
X = t=l
i=l
:
n.
Eci^t F = ^ n. where JloiiAi r,
z =
1=1
n. A We need data on the dependence of yield capacity of different grades on the amount of fertilizers and investments to carry out land improvement. Based on these data, we have to construct production functions and calculate optimal yielding capacity which enters into the grain farm optimization problem as coefficients. The above problem for the two-dimensional case can be solved by the method of dynamic programming. By employing the models for optimal allocation of major funds and for that of labor (this includes training), we determine major funds and the operating time fund for the grain farming. Prom the optimization model for foreign trade relations we can derive constraints on grain import. An important constituent of the grain chain is the problem of mixed feed manufacture optimization. The present text does not discuss the problem in detail. §33. G R A I N F A R M I N G OPTIMIZATION PROBLEM §33.1. The problem of grain farming optimization is a dominant link of the grain chain. The solution of the problem gives the possibility of obtaining information on the relation of home grain production and grain import, the optimal satisfaction of needs, the cost account of preservation, drying and storage of grain, the investments needed to achieve a particular level of
246
Systems
Optimization
Methodology
production and the efficiency with which they are utilized. The problem of grain farming optimization represents the deterministic problem of nonlin ear programming with 84 variables and 74 constraints. It must be solved by 3 objective functions for the long-range plan period. Its solution can be utilized in developing the final plan for grain farming. The model was constructed under the following assumptions: (1) The amount of grain in the state reserve does not change. Each year the state reserve is replenished with new grain, therefore the problem does not involve the variable which denotes the grain intended for the state reserve. (2) The self-cost for 1 ha of the area sown with grain includes the expenses for transportation of grain to a reception center (5 km). Transportation network optimization is the problem of local planning bodies. (3) Harvesting continues for 20 days. Grain arrives uniformly at a reception center. (4) The model is calculated for the long-range plan period. New construc tion and capacities obtained due to employment of investments imme diately come into service. (5) 10 percent of the entire grain harvest do not need drying. Water content in grain does not exceed 14 percent. §33.2. Objective Functions. We take the following as objective functions: (1) maximum profits; (2) minimum self-cost; (3) maximum domestic production. Maximum profits: At present, one of the national economic development criteria is assumed to be the national income increment, and to isolate the part of the planned increment of the national income which accounts for grain and its products is extremely difficult. Profit is the economic category close to that of the
Application
of Solution
Techniques for Large Dimension
Problems to . . .
247
national income increment. It is this category which has been taken to estimate grain farming and in the objective function we estimate by positive coefficients complex: food products, import grain, and fodder grain. The profit obtained at the earlier processing stages is proportionally included in these coefficients. Minimum self-cost: In this objective function, all variables are estimated. Of special inter est here is the cost development of the complex intended for preservation, drying and storage of grain. Maximum domestic production: The main problem of the national economy consists in providing the country with an adequate quantity of raw material. Domestic production of grain is insufficient to satisfy the needs. Of vital importance is the knowledge of the quantity of grain the farming industry can produce, dry and store, and available investments (though it may be useless in the aspect of maximum profits and minimum self-cost). §33.3. Variables. The problem employs 84 variables, out of which 82 variables have eco nomic meaning and 2 variables are auxiliary, assuming the value to be either Oor 1. All variables are subdivided into several groups. The first group deter mines grain production by individual cultures on lands of different quality, since the same culture yields differently on different lands and requires, accordingly, different expenditures of operating and machine time, etc. By convention we consider 3 kinds of land: — good land — medium land — bad land. The entire land is estimated by a hundred-mark grading system. The best land obtains 100 marks. Each plot of land is graded by the following data:
248
Systems Optimization
Methodology
— soil composition — geological origin — soil condition (e.g. slow changes in soil such as dewatering, decrease of lime content in soil composition) — local climate — characteristics of the place — local hydrological conditions and it is graded from 1 to 100 marks. Good land is assumed to be that with 60 - 100 marks, medium with 30 - 60 marks and bad land is credited with 1 - 3 0 marks. It should be noted here that agricultural treatment is worthwhile on the plots of land estimated with more than 16 marks. For each culture it is advisable to till only two kinds of land: either good and medium or medium and bad. It is unjustified to cultivate, say, wheat on bad land, where its yield is poorer than that of rye, and, besides, it will require more expenditure. A unit of measurement for these variables is a hectare. We then get information on how many hectares of land that should be sown with a specific culture. A full list of variables is given below. ii9 is the straw obtained from grain production; 120 is the portion of straw left in the field. By employing this technology (leaving a portion of straw in the field), we reduce the expenditure of working and machine time per 1 ha of the plot sown with grain. Variables a ^ o - ^ denote the tons of different grades of grain imported from developing and developed countries. The following group of variables determines the capacities of preserva tion, drying and storage employed out of those available at the beginning of the last year of the long-range plan period. Apart from preservation in the open air and drying, the preservation capacity features only technical devices. The area needed for it, enters into the storage capacity which is vacant at that moment. The units of measure for the variables denoting preservation and storage are exact and those for the variables which denote drying in the units of different types are exact for a 20-hour day.
Application
of Solution
Techniques for Large Dimension
Problems to . . .
249
The drying capacity available at sugar mills are not designed for grain drying. It is held in reserve when, due to poor weather conditions, we have to dry during harvesting an extremely large amount of grain with high water content. The group of variables determines the amount of individual food prod ucts made in the last year of the long-range plan period. Flour manufac turing involves two technologies of processing grain: in 12-stone mills (a routine practice) and in 6-stone mills. In the latter case the expenditure of working and machine time are taken into account. This is identified as the rapid milling procedure, the introduction of which requires large in vestments. The problem also contains such variables as the rye flour made by standard milling procedure and the rye flour made by rapid milling procedure. The unit of measure for rye and wheat alcohol is a hectoliter, while for other food products - tons. The following group of variables represents fodder grain, which is sub divided according to cultures, and by-products of the food industry also intended for fodder. The units of measure used are tons. A large group of variables denotes the new capacities of preservation, drying, storage, reception, cleaning and machinery needed for a continuous process of grain reproduction and those available due to employment of in vestments. For every variable we take into account the existing three kinds of investments, namely, investments for an increase in the available capac ity, i.e. for new construction works, investment for expanding the available capacity and those for reconstruction. At the cost of the first kind of in vestments we obtain the capacity for simple preservation, air supercooling preservation, drying in portable units and silage towers with the carrying capacity of 40 tons per hour, reception, cleaning, outdoor preservation, stor age in granaries, in a complex of silage towers, in silage towers (80 thousand tons), in silage towers, (with the capacity of 38 thousand tons), designated by appropriate variables. At the cost of the second kind of investments we obtain the capacity for drying grain in dryers (32 t / h ) , for storing grain in silage towers (10-20 thousand tons) and their annexes. The ton is taken as the unit of measure of the capacity for preservation and storage, the tons per day are defined as the units of measure of the
250
Systems
Optimization
Methodology
capacity for reception, cleaning and drying, and the hours are the units of measure of the machine time fund for tractors, lorries and combines. The last group of variables designates the obsolete capacity of flour milling by the standard procedure, the annual fund of machine time for obsolete tractors, lorries and combines, and the period of implementation of the long-range plan. In a sense these variables express obsolescence. By way of example, we consider the combines. Let the annual fund of machine time for the combines that have been in service since the beginning of the long-range plan period be, say 1000 hours. The planned number of the combines worn-out at the moment has been already subtracted. The time expenditure norm for old combines per 1 hectare is o. By employing investments, we obtain the auxiliary fund of machine time for new combines XH which are, say twice as productive, i.e. their norm of time expenditure for 1 ha is a/2. Let the number of hectares be X, then the limitation on employment of the machine time fund for combines may be written as follows: aX < 1000 + 2xH , or aX -2xH
< 1000.
The available time fund (1000 hours) is not utilized in full, since from the standpoint of self-cost, profit and working time expenditures it is more useful to employ new combines, though they are associated with investment expenditures. The old combines are idle, they are obsolescent, but we have to pay for them. Because of this, we include the variable xc denoting the machine time of the old combines which are not utilized due to obsolescence. The variable xc enters into the objective function profit maximum with a negative coefficient of magnitude equal to the payment for funds calculated for 1 hour. This variable also enters into the objective function - selfcost minimum with the same magnitude, but positive sign. The limitation on employment of the machine time fund for combines then has to be written as aX - 2xH + xc = 1000.
Application of Solution Techniques for Large Dimension Problems to . . .
List of Variables wheat on good land x2 - winter wheat on medium land x3 - spring wheat on good land x4 - spring wheat on medium land x 5 - winter rye on medium land x6 - winter rye on poor land x7 - spring rye on medium land X E - spring rye on poor land xg - winter barley on good land x l o - winter barley on medium land x l l - spring barley on good land x l 2 - spring barley on medium land 2 1 3 - brewer's spring barley on good land a 1 4 - brewer's spring barley on medium land 2 1 5 - oat on medium land 2 1 6 - oat on poor land 5 1 7 - mixture on medium land xis - mixture on poor land 2 1 9 - (total amount of) straw xzO - the straw left in the field XI
- winter
I1
x 2 ~- L'durumwheat" 2 2 2 - wheat 2 2 3 - barley 2 2 4 - maize 5 2 5 - oat
maize xz7 - millet - oat x26 -
imports from developing countries
imports from developed countries
barn storage 2 3 0 - granary storage xsl - storage in a complex of silage towers xz9
-
i
available capacity
251
Systems Optimization Methodology
oaten products - barley flour products - flour made from durum wheat - by standard milling procedure;
232 533 234 235
\
made from durum wheat by rapid milling procedure; 5 3 7 - rye alcohol 2 3 8 - wheat alcohol 5 3 9 - wheat malt zdo - brewer's barley malt 5 4 1 - barley treated for brewing 5 4 2 - maize starch 2 4 3 - wheat starch 2 4 4 - oat 2 4 5 - brewer's barley 2 4 6 - barley $36
- flour
rye - wheat - millet - maize - wheat for feed meal - rye for feed meal - flour product manufacture waste
247 248 $49 250 251 252 553
rye bran x 5 ~- wheat bran 2 5 6 - maize processing waste 5 5 7 - reception capacity 254 -
cleaning capacity - granary storage
258 559
I
+
products of food industry
1
IJ
grain and secondary products for feeding
grain and secondary products for feeding
Application of Solution Techniques for Large Dimension Problems to . . .
£eo - storage in a complex of silage towers £ei - storage in silage towers, capacity 80 thousand tons £62 - storage in silage towers, capacity 38 thousand tons £63 ~ storage in silage towers, capacity 10-20 thousand tons z64 - storage in silage tower annexes £55 ~ machine time fund for tractors and lorries Xee ~ machine time fund for combines £57 - milling capacity (standard milling procedure) £68 - machine time fund for outdated combines £59 - machine time fund for tractors and lorries £70 - drying in stationary driers £7i - drying in portable driers £72 - drying in driers (32 ton/h) £73 £74 £75 £76 £77 £78 £79
-
253
capacity obtained due to investment unitilization
outdated capacity
available capacity
drying in portable driers drying in silage towers (40 ton/h) preservation by air supercooling (new capacity) simple preservation (available capacity) simple preservation (new capacity) outdoor preservation (available capacity) outdoor preservation (new capacity)
new capacity
§33.4. Constraints and coefficients.
The problem has 74 constraints: agricultural, constraints for grain cleaning, preservation and treatment, general economic constraints and constraints, which characterize satisfaction of needs for food products and grain fodder.
254
Systems
Optimization
Methodology
The first 6 constraints bound above and below the sown area are subdivided into good, medium and poor land. Constant terms of these inequalities can be obtained from the models for optimal allocation of major funds. The lower bound for each kind of land is employed to compel farms to sow a particular area of each kind of land (which is vital to crop rotation) and thereby achieve a certain level of production. These constraints play an important part in solving the problems for maximum profit and minimum self-cost. Thus we have the following inequalities: x\ + Xi + Xx33 + + xg Xg + iXUn + + x X13 < bi bi , 13 < X\ + X\ + XX33 + + Xg Xg ++XXn n ++ X13 X13>> 62 62, s x% ++X4+ X-2 £4 4-X£5 S+
+ X£7 £10 X12++ Xx114 + X^15 xn << 63 63, , 7 +4-X 1 0 +-I-X12 4 -I1 5 ++ X17
12 + 12 + £4 14 + + X5 £5 ++ 17 17 ++ £10 £10 ++ Zl2 &12++I lIl4 4 ++^15 *1S++Zl7 Zl7>>64k(1 1 16 +^8 X& + X 8 +*16 + * 1 6 +Z18 + Z 1 8 << hh , , XQ XQ
+XX$8 + +Xx1iS8 >> &6 66 .• + + x Xi61 6 +
Arable land in the country is limited, therefore we have to be careful not to allow separate plots to become lixiviated by the same crop. For this purpose, each farm needs to arrange for crop rotation, which is an advantageous succession in cultivating different crops on separate plots of the sown area over a particular number of years with the object of achieving a constant high yield while preserving fertility. Utilization, preservation and increase of soil fertility are determined by rotating crops on plots yearly. In this aspect it might be useful to cultivate in one year only one crop at each farm. But it is impossible from the economic point of view, i.e. from the standpoint of utilizing basic and current assets as well as the working time fund since the work to be done by all participants in production would be unequal. When constructing crop succession we need to comply with the following: (1) Accommodation of a plant to itself and other plants. If the plants are accommodating, then we do not observe a drop in yield; if they are not accommodating, then there is a danger of pest accumulation. (2) We should rotate crops which either improve or deteriorate soil structure.
Application of Solution Techniques for Large Dimension Problems to . . .
255
(3) We should choose, for a plot, crops which secure an increase in the yield of the subsequent culture. For individual grain crops, we have, say, the following data: useful rotation:
winter wheat oat winter wheat winter rye
- oat - winter wheat - winter rye - winter wheat
possible rotation:
winter wheat winter wheat oat winter rye
-
impossible rotation:
spring barley winter barley spring barley
- winter barley - spring barley - winter wheat
winter barley spring barley winter barley winter wheat
Crop rotation cannot be confined to any one farm, it has to be carried out throughout the country. Therefore the problem involves crop rotation limitations on the amount of the cultivated wheat, barley, oat, spring grain as a whole and spring barley. Here we take into account the above accommodation of grain crops to themselves and other agricultural crops. For example, the number of hectares sown with wheat and barley must not exceed 65% of the entire area sown with grain. The exact percentage is obtained from the model of optimal allocation of the basic agricultural assets. The constraints for imports are derived from the model of foreign relations optimization. On the one hand, the amount of grain imported from developing countries is bounded from above and, on the other, the same is with the sum of currency roubles for purchasing grain in developed countries. The coefficients with variables £ 2 6> £27, ^28 are equal to the price of 1 ton of appropriate grain grade in currency roubles. Constraints 13 and 14 are related to straw. Equation (13) evaluates the amount of straw to be produced in the planned year. We know the planned yield of each crop on different lands as well as the planned ratio, grain:
256
Systems
Optimization
Methodology
straw, which is, say for winter wheat
1 :: 1.2 1.2
for spring wheat
11:1.1 : 1.1
for winter rye
1 :: 1.4, 1.4, etc. etc.
If, say, the planned yield of winter rye, which is cleared from the harvesting and drying losses, on medium land is 3 tons, then the planned yield of winter rye straw on medium land is 4.35 tons. Then we have equation (13): 18
2jai3;£j 2 j a i 3 j Z j - xig = 0. Inequality 14 secures satisfaction of the requirements in the planned year, in the straw intended for litter and industrial needs, i.e. ££19 l 9 - ^20 £20 > hi ■
&14 will become known to us from the fodder balance and the balance of light industry. Equations (15)-(26) secure satisfaction of the needs for products of food industry, i.e. the needs for flour products made from oat and barley, for du rum wheat flour (which exhibits especially high quality), flour from wheat and rye, alcohol from rye and wheat, malt from wheat and barley for brew ing, brewer's treated barley, starch from maize and wheat. The needs of the population for grain food products remain nearly the same for subse quent years. The grain, from which these products were made, might have been more efficiently fed to livestock and used to satisfy the needs of the population. Therefore all constraints with the exception of those for flour manufacture are given say, as the equality: £32 = h5.• £32 = &15 Now these x's have ceased to be variables. They are fixed to one value. We can eliminate them and thereby considerably diminish the problem's dimensions by evaluating through the use of the coefficients of material expenditure the requirements in separate grain grades for the food industry.
Application
of Solution Techniques for Large Dimension
Problems to . . .
257
The amount of secondary products obtained from production and intended for fodder, are readily determined by coefficients. The problem retains these variables since it is the main link of the grain chain, and the grain chain may provide an example for constructing other chains. Subsequently, we need to diminish the dimensions of the problem, where the grain for the needs of the food industry will enter into appropriate equations as a constant term. Accomplishment of constraint 27 secures satisfaction of the needs for fodder grain obtained from the fodder balance. It takes the following form 2.86x7 + 2.17x8 + 3.16^17 2.52xi 8 + 0.72xi 3 + 0.67x14 2.86x7 3.16x17 + 2.52xi8 + X47 x 4 7 ++ xx 44 88 ++ X49 149 ++ £50 x 50 ++ X51 x 5 J >> 627 627 •• + The needs for fodder maize and millet will be obtained from the fodder balance separately. We have therefore the inequalities 152 > 152 > &28 &28 1 1 Z53 > &29 ■
It may be said that one fifth of the brewer's spring barley cultivated, does not exhibit the properties needed for further processing. This portion of barley is used for feeding. The planned yield (cleared from losses) of this special spring barley on good land is 3.85 tons and on medium land, 3.6 tons. If we subtract from this yield the needs for seeds per 1 ha for the next year (which is on the average 0.2 t) and the needs for 35% of storage losses, then the "pure yield" is 3.6 tons on good land and 3.35 tons on medium land 3.35 3.6 =0.67. — =0.75, = 0.75, T 5 5 5 It is assumed that storage losses account for 1.5% of the stored grain. But during harvesting and in the process of drying, 1/12 of the annual output is consumed, i.e. only 11/12 of the entire grain output is stored for some time, and this accounts for 1.5% of storage losses. Thus it may be stated that storage losses of the entire output are 1.35% « 1 . 5 % - ^ % . 1z
258
Systems Optimization
Methodology
In equation (27), x 49 denotes the amount of brewer's spring barley, which, apart from these 20%, is used for feeding. Spring rye is cultivated only when winter rye is destroyed by frost. It is completely used for feeding. The rye is incorporated in the problem, since the weather conditions for the planned year are not known, i.e. we do not know whether we have to cultivate this rye or not. Evidently, we need then to incorporate in the problem the additional constraints on the area sown with spring rye on medium and poor land, which should not exceed a particular percentage of the area sown with winter rye: 17 — — 01X5 a i i s << 00 ,,
x8 - axx6 < < 0. Mixture is completely used for feeding, too. Therefore, spring rye and mixture can be directly included in the fodder grain equation in the same manner as 20% of a brewer's spring barley. Equations (32)-(37) are the balance equations for the secondary pro ducts of food industry. Thus we have: O.IX34 + O.IX35 0.1135 + 00.1X36 . 1 l 3 6 + O.IX37 0.1*87 - X54 = 0 ,
i.e. we obtain 0.1 t of feed-grade flour wheat from the manufacture of 1 t of flour from wheat. These constraints fall out of the problem in the above reduction of the problem dimension. Constraints (38)-(45) are the balance equations for the manufacture and allocation of wheat, rye, durum wheat, barley, brewer's barley, oat, millet and maize; thus wheat provides an example of such equations: 4.05xi + 3.16x2 + 3.95x3 + 3.06x - l-33i3 1.33x366 - 1.33x37 3.O6144 + x 22 ~ —
— 0.26X41 — 0.26X41 ~~ 1-238X42 1-238X42— 2X46 2X46 ~~ £51 £51 == 00.. Here the coefficients with the variables designating grain production are equal to the planned yield cleared from the losses of yield and drying and 1.35% of storage losses, and the needs for seeds per 1 ha. The remaining coefficients are the coefficients of material expenditure for producing 1 ton of separate food products. Inasmuch as the needs for wheat alcohol (let it be 1000 hectoliters), wheat malt (say, 1000 tons) and wheat starch (let it be 2000 tons) are known, the constraint is transformed into
Application
of Solution
Techniques for Large Dimension
Problems to . . .
259
4.05*1 + 3.16x2 + 3.95x3 + 3.06x4 + x 22 - 133£ 36 - 1.33x37 - *si = 5506 , i.e. grain for the needs of the food industry with the exception of flour manufacture shows itself in the balance equation as a constant term. It is the same with the balance equations for rye, barley, brewer's barley, oat and maize (durum wheat is milled into flour, millet is used for feeding). Equations (46)-(52) involve the upper restrictions of the capacity being available at the beginning of the last year of the long-range plan, for simple preservation, drying in stationary driers (tons per day), drying in portable driers (tons per day), storing in barns, granaries and in a complex of silage towers and for outdoor preservation. These equations are obtained from the capacity available at the beginning of the long-range plan minus physical wear of the capacity for the long-range plan period. Thus, for the problem we took the capacities available at the end of the preceeding period less the planned physical wear in the earlier years (here the capacity of driers has already been recalculated for 8% of dewatering). Assume that the harvesting period is 20 days and let, say, 66.6% of the grain output be sold to the state. We need to clean the grain and dry the remaining 90%, i.e. about 60% of the entire yield. The remaining 33.4% are left and treated at farms. Their treatment is not centrally planned. Assume that harvesting proceeds uniformly during these 20 days, then we have to process 1/20 of the entire yield each day. We denote the variables by E
60% of the entire grain yield
b
the capacity for drying per day
Vi
the capacity for preservation by air supercooling
V2
the capacity for simple preservation
V3
the capacity for outdoor preservation.
Then for each day we must have 20
20
20
20'
As stated above, in the case of preservation by air supercooling the grain, storage time is 30 days, and when the time is up the grain dies. The grain thus preserved does not die during harvesting. Each day we observe an
260
Systems
Optimization
Methodology
increase precisely by 1/20 of the entire capacity Vj (It may be assumed that an increase in each kind of preservation is uniform, since, even though the cost structure changes each day, it remains the same as a whole). The second preservation technique surveys the grain storage time being 14 days before the grain dies. In this case, grain uniformly increases during the 14 days. After this period we have to dry the grain which has arrived for preservation on the first day. We release thus 1/14 of the capacity having been employed until now or 1/20 of the entire capacity V2 which is replenished in its turn with the grain harvested on the 15th day. On this day a double amount of grain arrives for preservation, since the drying capacity has actually decreased by V2/2O. The same is with outdoor preservation, though in this case the grain storage time before drying is 7 days at most. To illustrate, we take the following example. Let the harvesting period be 14 days (T = 14). Thus we have one preservation technique. In this case the maximal time of grain storage is 5 days (c = 5). Let
E = 532 tons b6 = 28 tons per day E1
then
— = 38 tons 14 14 1? and b — 38 tons — 28 tons = 10 tons 14 V should be 140 tons or 10 tons. On the sixth day 38 tons of the harvested grain arrive again. But only 18 tons out of these 38 tons can be dried, since on this day we have to dry those 10 tons which have arrived for preservation on the first day. Twenty tons of the incoming grain are preserved, 10 tons of the total amount (= V/14) occupy an empty place and another 10 tons (= V/14) take a new place. Thus, for the sixth day we also have the equality — - b - — 14 ~ 14 ' 14 ~ 14 ' Furthermore, we need to secure the ratios provided that the grain stored by the outdoor preservation technique dries if only for 7 days V V3 <<7b, 76, 3
the grain stored by simple preservation dries for 14 days: V V 2 << 146 146 2
Application
of Solution Techniques for Large Dimension
Problems to . . .
261
and the grain generally stored by preservation dries for 30 days. Vi + V2 + V3 < 306. 306.
The minimal drying time is 20 days, and the maximal drying time is 50 days (20 + 30). The real drying time obtained from the problem is equal to the ratio E/b, i.e. the overall drying capacity accounts for {E/b) ■ b. The self-cost of drying 1 ton of grain in different driers is not the same and the self-cost of drying the overall amount of grain which constitutes thereby 18
djXj 772J djij j=i v^ j=l V"^ 77 77
c x 2—i 2—1 iciixi ' '
E *i >=73
3j = 73 73
where dj = 0.9d'j and d'j is 66.6% of the planned amount (cleared from harvesting losses) of the demand for seeds per 1 hectare of the area. We determine the self-cost of preservation of the entire amount of grain. The self-cost of preserving a ton of dry grain for 1 day by different tech niques and in different devices is known, it accounts for & {j — 78, 79). Hence (see the plot) the self-cost of preservation is equal to 19
10CQ + IOCQ + 220c[, 04 + + ■■ ■ ■■ ++ 130CQ 130CQ++ 140 1404 + 11124 1 2 4 ++ 84c' 84c'0 0++56CQ 56CQ++28c' 28c' —■140 140• •c'c' C; + 0 0== —■ 00
or, in a general form, it amounts to: c '
0
- V - ^
= coV(T + c),
where c0 = c(,/2, though T + c = E/b, is the drying time. The problem involves 3 preservation techniques, namely, for preserva tion by air supercooling T + c = 50, for simple preservation T + c = 34, for outdoor preservation T + c = 27, i.e. these capacities can be best utilized for 49, 33 and 26 days, respectively. The real time of their utilization depends on the magnitude 18
E dixi j=i j=73
„ _ £1
262
Systems Optimization Methodology
The first kind of preservation capacity is employed in any case for E/b 1 day. If 34 < E/b < 50, then the second kind of preservation capacity is employed for 33 days and the third kind of preservation capacity for 26 days. If 27 < E/b < 34, the second kind of preservation capacity is employed for E/b - 1 day, and the third kind for 26 days. If 20 < E/b < 27, then the preservation capacity of the second and third kind is employed for E/b - I day.
Application
of Solution
Techniques for Large Dimension
Problems to . . .
263
Considering c, = c'j/2 (j - 78,79), we obtain that the self-cost of preser vation accounts for is is d x E £ d33Xi3 — —
rr
is is ££ d3diXx3i 34s 34S + + (1 ( 1 -~ s)~ S ) ~
cC7S7SxX7&7& + +
E £ X*j 3
3> = 73
+ +
|_
J 27t 27t + + (l-t) (l-t)J~ ~
cC78 X78 78x 78
££ X^i 3
3=73 J=73 18 2_, 2_, ctjXj ctjXj
cc79xx79,, 79 79 £E
j=73 3=73
X
*j 3
where 18
E •***>
uu
E
d X
33
34 34 + + £t*$ E »J
34s + ( l - s ) ^ < £E *i .7 = 73 7=73
p—;
18 18
E ^***
18 18
E 4?*j
27t + (1 - *)*§ X £ *$ 3 3£ = 73
27+^ x E x> '
<
p _
j=73
It is to be noted that preservation by air supercooling can proceed only in the silage towers with the capacity of 18 thousand tons and 38 thousand tons and simple preservation in granaries and in a complex of silage towers, x < xX64 #78 < &5 64 + X e5 ,,
X79 < + 131 X31 + + ^62 X62 + #63 •■ < X £30 30 +
Furthermore, let it be planned that in that year if only 30% of the entire preservation capacity must be the capacity for preservation by air super cooling, then Vx > 0.3(Vi + V2 + V3).
264
Systems Optimization Methodology
T h e point is that if the weather conditions are poor throughout the harvest ing period, then a large amount of humid grain has thus to be dried and preserved. Simple preservation then produces the desired effect, because the air by which grain is aerated exhibits a high percentage of humidity. T h e silage towers with the capacity of 10-20 thousand tons and their annexes are added to the available complexes of silage towers. But this is not true for each place. Thus in some regions, where grain is little culti vated, a sufficient amount of storage capacities is available. T h e problem, therefore, includes the restriction t h a t for the long-range plan period the silage towers (10-20 thousand tons) or their annexes can be added to not more than 30% of the available capacity for storing in complexes of silage towers. The number of these complexes of silage towers is 0.3x11 0.3x3i
(their (their capacity capacity is is 8300 8300 tt oo nn ss )) ..
— —I~ ^ « 0.00004x 31 8300 « 0.00004x3!
The number of annexes to silage towers is
nocx,* 67 * 0 0 0 0 1 X 6 7 T h e number of silage towers (10-20 thousand tons):
15000
Xm X 0.00007X66 66 .■ 66 » 0.00007:r
Let us assume that, on average, 1.5 annexes or silage towers (10-20 thou sand tons) can be added to every complex of silage towers. Constraint 60 thus takes the form: 0.00006x31 > 0.0001x 6 7 + 0.00007x66 . Constraint 61 is the balance equation for storage 18
28
a x
31
67
Y ^ 3 + g Y 3■■i - Y Y*s*i - ~ Y 5Z ®3xi --° °- ■ j=l
m x
j=21
:
j=29
3j == 6 2
In this case we assume that during 2 months of grain harvesting and drying, 1/6 of the annual imports arrives for storage. T h e ratios Oetj constitute 11/12, i.e. 66.6% of the planned amount of grain (cleared from losses).
.Application of Solution
Techniques for Large Dimension
Problems to . . .
265
One twelfth of the entire yield is fed to livestock or processed by the food industry during harvesting. Equation (62) restricts employment of the milling capacity (standard milling procedure) measured in machine-hours for the beginning of the last year of the long-range plan (the planned capacity for the beginning of the planned year minus physical wear). Let the data on this point be obtained from the forecast, thus we have b62 = O.667X34 0.667x36 + 0.667x38 + x 70 . 0.667x34 + 0.667136 Constraint 63 bounds from below the milling capacity (rapid milling procedure). The constraint allows a particular percentage of the milling capacity (standard milling procedure) to be reconstructed for the rapid milling procedure over the long-range plan period: b63 < 0.3x35 + 0.3x 37 + 0.3x39 . The machine time fund &63 calculated for the standard milling procedure does not naturally enter into b^The constraints on other capacities for making the remaining grain food products are not included in the problem. These constraints, much like the appropriate constraints on the machine time fund, enter into the models optimizing the satisfaction of consumers (the model optimizing the satis faction of consumer's needs for flour is not constructed). Equality (64) yields a restriction on the machine time fund for tractors and lorries 18 18 &64 = /l*9*jxj ~ a 64,20£20 - l - 5 x 6 8 + X 7 2 , 664 = }a64jXj - a64,20£20 - l-5x68 + £72 > J= l
where &64 is the machine time fund of tractors and lorries for a grain farm, which is available at the beginning of the first year of long-range plan minus physical wear before the last year of the long-range plan. The magnitude is obtained from the forecast: are the coefficients of machine time expen ditures for old tractors and lorries (per 1 hectare) during a year. This includes all expenditures for preparing machines, for sowing, fertilizing and harvesting (apart from combines). 064,20 is the amount of machine time saved by leaving 1 ton of straw in the field.
266
Systems Optimization Methodology Equality (65) expresses a restriction on the machine time for combines 18
65ix3 ~ - 2x 6 9 + Xn Xn , &65 = 2_] a65j iJ == il where 665 is the machine time fund for old combines which had not been physically worn out before the last year of the long-range plan. T h e number of these combines is obtained from the forecast. We obtain the machine time fund by multiplying the number of combines by the harvesting time and the operating time of these combines. The variables X68 and rr69 in the objective function of the self-cost min imum are not estimated by positive coefficients. These expenditures enter into the planned self-cost of 1 hectare sown with a particular crop. De termination of these coefficients, particularly for different kinds of land, presents a serious problem. We should take into consideration the changes in technologies, the amount of mineral fertilizers applied, the changing ex penditures of the direct and embodied labor, etc. Each day 1/20 of the yield (equivalent to 66.6% of the overall o u t p u t ) arrives at the reception center. Restrictions for the reception capacity are given by inequality (66): 1 18 he > So 2 ^ 4 * 3
_I
60.
J=I
(The reception capacity is the grain carrying capacity per day at the re ception center. It is determined by the equipment of the center, namely, scales, conveyer systems, etc.) b66 is obtained from the forecast. It should be noted here t h a t the present model does not consider special treatment of seeds. It takes into account only the place needed for seed storage. The amount of grain arriving every day could be expressed also in terms of the daily output of combines: hs n . n .. fa l-"2m d -' °° * - 55
+
X69 ^0 "20
X7i ■1-0-a--2o--0.5-aj.0.666,
where 0.5 and 1.0 represent the number of hectares to be harvested for 1 hour by appropriate combines. But the present formula includes a the
Application
of Solution
Techniques for Large Dimension
Problems to . . .
267
average planned yield which seems to differ from that obtained when solving the model. This may cause an error in the arrival of grain per day up to 150 000 tons. In a sense, the outdoor preservation capacity may be thought of as a reception capacity, since the grain is actually "received" only when it is intended for drying, thus the equation includes the term. The grain received is immediately cleaned, which means that the clean ing capacity must be, on one hand, larger than the reception capacity or equal to it and, on the other, 90% of the cleaning capacity must be larger or equal to the amount of grain which is dried or preserved each day by two procedures (to provide synchronism between receiving, cleaning, drying and preserving), i.e. &67 > #XQO 60 — — XQI
,
where b^j = c\ — &66, c\ is the cleaning capacity available at the beginning of the last year of the long-range plan; and 0.9(Cl+i61)> or b&68 = 0.9ci 0.9ci > > 68 =
Vi1 + V ¥i
V +V Vi v2 2Q 2Q
2Q
+b,
+ bb -- 0.9x 0.9x6ii •• + 6
The variables xso and XQI in the objective function - self-cost minimum are not estimated. The self-cost of receiving and cleaning 1 ton of grain does not enter, as a constituent, into the production cost of 1 ton of grain. We consider now a restriction on investments beg as the available sum of investment for a grain farm. By subtracting from this sum the capital outlays for land improvement at the farm, it is obtained from general na tional economic restrictions for the given period of the long-range plan. As the coefficients with the variables for flour production by the rapid milling procedure we assumed the weighted sum of capital outlay for producing 1 ton of flour by the reconstructed capacity of three kinds of facilities with outputs of 31.5 tons/day, 63 tons/day, and 125 tons/day. Weights repre sent the planned percentage relation of their establishment. The coefficients with x60, xBi and x75 - x77 denote the sum of capital outlay needed for construction of a new capacity for processing 1 ton of grain per day (i.e. receiving, cleaning, drying). The coefficients with the variables x62 - x67
268
Systems Optimization Methodology
and i 7 8 represent the sum of capital outlay needed for the construction of a new capacity for processing 1 ton of grain. The coefficients with i68 and rE69 in the equation denote the sum of capital outlay needed for an increase in the annual machine time fund by 1 hour. The latter three equations are the restrictions on the annual fund of operating time in the farming industry itself, intended for grain production, flour industry, preservation, drying and storage. We first consider a constraint on the operating time fund for the farming industry: 18 18 &70 > V , a70jXj
— °-x20 _ ^ 6 8 — ^ 6 9 i
= ll jJ =
where aroj are the summary expenditures of the operating time for 1 hectare sown with grain, considering that old technology and old machinery are employed. When employing the technology, according to which straw is left in the field, we save a hours for 1 ton of the straw left in the field. When employing new machinery, we save a hours for every hour of utilization of new tractors and lorries, and employment of combines saves 6 hours. Exact evaluation of those coefficients (a, a, a) is particularly difficult, since along with the employment of new facilities we also employ new technology, which cannot be included in the coefficients representing the average value for the entire farming industry. Evidently, the model here needs careful revision, particularly from the economic point of view. Constraint 71: 39 39
&71 > ^2 a71jXJ -' j = 34
where a-nj denotes the operating time expenditures for producing 1 ton of flour by different milling techniques. Below is given the formula, where nonlinearity makes its appearance again just as in the objective function - self-cost minimum due to the variables for drying and preservation. The coefficients a72j, j = 7 3 , . . . ,79, are the expenditures of operating time for 0.5 ton of the grain being dried and preserved in different facili ties. The remaining positive coefficients a-jij for j = 29,30,31,32,... ,67
Application of Solution Techniques for Large Dimension Problems to . . .
269
represent the expenditures of operating time for storage of 1 ton of grain over 6 months, since utilization of grain for feeding and industrial needs is assumed to be uniform throughout the year. Therefore, on average, every ton of grain is not stored for more than 6 months. We have 18 72
Z ^ OjXj
a
X
b72 > J2 " j J + ^ 7 i=l
£
j=73
a72
5Z
^
Xj i=™ 18 / J / J
+ +
TS
345 34s+ + (1 ( 1- -S5 )) ~^
E E
J = 73
djXj &jXj
X X
aa72 ,7 8 X78 72 ,7 8 Z78
••
J3
It should be noted that the present inequality involves double counting, since part of the operating time for drying and preservation enters into the operating time for storage, and to what extent it happens is not known for the present. To be more precise those expenditures are included in the two inequalities given below. But, on the other hand, definition of constant terms presents a serious problem. The expenditures of operating time for receiving and cleaning enter into the coefficients of the operating time expenditures for producing 1 ton of grain. &701 hi i and 672 are derived from the model of optimal distribution of the operating time fund of the farming industry. §33.5. We consider the objective function coefficients. In the function self-cost minimum, the coefficients are equal apart from the import of the obsolete capacity and the straw left in the field, the cost of processing, or the self-cost minus the expenditures for raw and grain materials (for import), are equal to the prevailing market price. The coefficient with x2o is equal to the decrease in the self-cost of producing 1 ton of grain from 1 ton of straw (due to the employment of a new technology). Fodders are not estimated here. In the function-maximum of profit, the coefficients with x2i - x28 are equal to a difference between the selling price of imported grain inside the
270
Systems
Optimization
Methodology
country and the market price. T h e coefficient with x2o is equal to C\Q with an inverse sign, though. T h e coefficients in the objective function - home production maximum are equal to the net output, or the planned output which is cleared from the losses of harvesting, drying and storing as well as from the needs for seeds per 1 hectare of the sown area. The variables Xig - X73 are not estimated in the objective function. §33.6. General formulation of the problem 79
^dijXj
^6,,
2i = 1 , . . . , 7 1 ,
J3==I1 18
72
/
E ^jXj djXj
^2a72jxJ
+ 3-^-^
^2a72jXj+ J 73
E
18
34s + ( l - s ) ^ =
=
\ 18 18
E di' *i
3 4 + ^
18
x J2 *> i 2 J2_j _ j
E ^ E *3j
j=73 j = 73
18
/ . djXj
27+^ 18
77
2_, djXj
art+ l-*)^—< m +((i-t)t±—< E E XXj J
2L, xj
if^; JI!_;
j = 73 j = 73
s = 0 , 1 ; t = 0 , 1 ; Xj * j > 00,, j = 11,,.... .. , 779. 9. 79
I:
max 2_.c)xj
i
072,79X79,
E
j = 73
E djXj 34s + (l-s)^7—< J
\
J2 d,Xj
7 g 78
j=73
/
Application
of Solution
Techniques for Large Dimension
Problems to . . .
271
18
72
12 djX]
78
II: rain I £ c]x3 + ^ Yl cc>3 Pi J=73 J 73 i=* Z *i *i = i=* 12 kIl
JJJ===73 73 73
+
34s + ( 1 - s ) ^
^ ^
\
\ 18 \
III:
% X]
J=73 J=73
072,79X79 > .
/I
i
//
JJ
max 18y^c^Xj max i=i yjcj^j •
§34. SOLUTION T E C H N I Q U E §34.1. We consider the problem
dn I Y c4*i min \%3 ++^3j=h j=l
I,^
J2 J2C)VJ fy] \\. '
12 % J=l J=l >4(X,y)>6, J=1
X Xj'j >> 0 ;
= l,ni ; j =
2/j > 0; s/j >Q;
ji = M 1,«2, $,
;
J
(34.1)
where the matrix A is the matrix of dimension m x (ni + n 2 ) and b is the vector (bi,b2,...
,bm).
The constraints on Y are such that 12 Vj > 0 in
any feasible program. We set forth the technique of solving the problem by parametric programming. A Parametric Form of the Problem Problem (34.1) is equivalent to a parametric problem, where a parame ter enters into the objective function and the coefficient matrix. Definition 34.1: Extremal problem (A,iF) and {A',F) (both for a mini mum), where A and A' are the sets of feasible programs and T and T' are
272
Systems
Optimization
Methodology
objective functions, are identified as equivalent, if there are the mappings 1
and i\) :A'-* A' ->A, A, it: satisfying the following properties: 1. T{a)>T'[
£ * J"
J
(34.1) (34.1)
A{X,Y) A(X,F) >b, >b, X>0, and
{
Y>0
"n 1i
W2 "2
i=i J=I A(X,r)>6, A(X,F)>6, dX-AeF =0, dX-AeF =0,
X>0,
x>o,
Y>0,
y>o,
where dd== (di,d ( d i , d2,... 2 ) . . . , d,d n ,ni)),, /=(!,!,...,1).
|
J
(34.2)
(342)
Application
of Solution
Techniques for Large Dimension
Problems to . . .
Problems (34.1) and (34.2) are equivalent. Proof: I. Problem (34.2) can be regarded as
{
"l
n2
I]
A(X,Y)>b, dX-\-X dX eY eY
x >o, x >o,
= 0,
yy >o, >o,
—oo < A < +oo —oo < A < +oo taking A here as a variable below the internal minimum we have min g(u, v) = min min g(u, v) = min min g(u, v). min min min U,V g(u, v) = min U V g(u, v) = min V u g(u, v). U
U,V
V
V
u
II. Suppose we have two problems (ft,/) and ( f t , / ' ) , where
rf(X,Y):A(X,Y)>b} (x,y):J4(x,y)>^ ft= I ft=I X>0 X>0 I\
yy>o >o
II ni
Z^ djXj
JJ nj
/(X F)== E£ 3** c^ , CJS + /(x y) +^ ^ — £E <&. J == 11
E Vj
i=ll
' ( x ' , y ',,A A))::^^((x ' , y ) > 6 ' ft = I
i
dX' - XeY' = o X' > 0
y >o - o o < A < +oo
/'(x,y,A) = E c X + *E c ^-
273
274
Systems
Optimization
Methodology
We need to demonstrate that there exist such mappings V :: nn -> ip -+ n' n' and and V V :: #$ -♦ -♦ R R exhibiting the properties
/(x.m/'Mxn). f(X,Y)>f'(v(X,Y)),
/'(x',r,A)>/(^(^',i",A)). f'(X',Y',X)>f{i'(X',Y'^))Let (X,Y) 6 fJ and ( X ' , y , A ) e SI'. As y>, we take the mapping
• ♦»afnti x n , ; j /jri. i , . .■ . , •y•n •irn«, 2 , A ) ,A),
where A is calculated from the formula V d,-x, J 3 J
. _ jjg
_ dX dX
Eij and, as $, we take the mapping V>(xi,..., << A)2,=A)(x[,..., ■ n2 y[,..., ). ) . #{*i. • ■ ■. ; l i■.., . ■ y' • n2 •.,V'n = ( x i , .x' . n2 •, x' ; y[,..., y'y'n2n2 a ;, y[, (1) We show that
ip O -> ', v ::fl -. 0o',
Let
V(X,r)eft, then
A(x,y)>6, X >0, V ( x i , • • •, xni;
t / i , . . . , ynt)
F>0, = ( * i , . . ■, x n i ; j / i , . . . , yn2, A ) ,
where X _
~W
dX
Application
of Solution
Techniques for Large Dimension
Problems to . . .
i.e.,
ddXx --XAeYe r == o0 and A e6 (—oo,+oo). Therefore,
V(ft') £ ft-
Let
V(X',r',A)eft\ then
A(x',y')>6, dJC'-Aey'=0, ff
x >o, x >o,
y>o, r'>0,
- o o < A < +oo, VK^i, • ■ ■, x'ni; j / i , . . . , y'n2, A) = (x[,..., Therefore, tf>{X',Y',A) (3) We demonstrate that
x'ni; y[,...,
y'n2).
eft.
/f(X,Y)>f{ ( X , y ) >V(X,Y)). />(X,Y)).
Let
v(x,y) eft, f{X,Y)=c'X+^Y, /(x,y) = c1jr + g c 2 F 1 1 2 f{
p(X,Y) = = (X,Y,\); (X,F,A); ( p(X,y)
bUt
Y
dX
275
276
Systems
Optimization
Methodology
Therefore l / ' (y(X, (tp(X, Y)) = c'X + Xc22YY = cc'X X + ^c2Y
= f(X,
Y).
(4) We demonstrate that f'(X',Y',X)>f{^(X\Y',X)). f(X' tY',\)>f(i>(X',Y',X)). Let
Y(X\Y\X)eil', V(X',F*,A)gO', f'(X',Y',X) /'(A",y',A) = = c11X' x' + + Ac Xc2Y', y, (X',Y'), {X,X),
^(X\Y',X) 4>{X\Y',X) = 1
f(^X',Y',X))=c X' /(v(x',r,A))=c but since
1
Ay i AY'
x' ++ ^ 7 ^c7c22rY\,
(X\Y',X)eQ', dX' -XeY' i.e. _ _ Therefore, f(iP(X',Y',\)) f(iP(X',Y',\))
rlX'
2 = = c'X' c'X' + + ^jC ^jC2Y' Y'
It is thereby demonstrated equivalent.
that
= 0,
dX> dX
'
l 2 = = cclX' X' + + Xc Xc2Y' Y' = =
problems
f'(X',Y',X). f'(X',Y',X).
(34.1) and (34.2) are
Considering that the problems are equivalent, both of them simultane ously have or do not have feasible programs and t h e image of an optimal program would be the optimal program in the equivalent problem. As indicated in the demonstration, t h e problems
{
m ni
jj=l =l
n22
jj=\ = l
A(X,Y)>b, A(X,Y)>b, dX -XeY = 0, dX -\eY = 0, X >0, Y >0, X >0, Y >0, —oo —oo < < A A< < +oo +00
j)I
J
Application
of Solution
Techniques for Large Dimension
and
{ X>0,
Y>0
"1 "1
Problems to . . .
™2
j
XS^c2.y.,
y^c)x};+ j=i
J=
277
>,
i
J
are equivalent. X > 0 , Y>0 By solving the internal parametric problem, we obtain the function
,
where the functions Xj(\) and yj(X) are linear fractional, in which the minimum of the internal problem is achieved. Let
{
rn
n2 712
71,.
jj
J2C)XJ+X^c^yj Y^C)XJ X^c)y3 A(X,Y) + = b, j=i j=i j=\
dX-XeY A(X,Y) X >0, dX-XeY
I\ , J
= 0, = b, F >0. = 0,
X >0, Y >0. §34.2. Definition of the Solution Stability Domain We denote by Ai the part of matrix A corresponding to X variables and by A2 the part of matrix A, corresponding to Y variables
A All
= =
^A22 = =
/ oi l
a
Wml Wml
a
<m22
••■ ••■
< &mnJ «,/
(/ 0?1 afi
aa
h
■■■
« li n 22
a \Wml ml
«m2
•••• •
a
i2 «12
••• •••
o4*1 } n i \\
,
a
\
mamn n 22/)
-.
278
Systems Optimization Methodology
In these designations the problem then takes the form min(c m i n ( c11A: X + Ac A c 2 F) y) * * + ^ = *> ddX-X x - eA Y=0, e y = 0, X>0, X>0,
(34.3)
Y>0. Y>0.
T h e dual problem to the above max for Im ■KA\ + + du> du <
where the dual estimate vector is represented as ( 7 r , w ) = (-Kl,TT T m m, w ) .. (7T,W) ( * ! , * 2*,., . .■. •, ,7 7T ,w)
We solve problem (34.3) with A = A 0 . When obtaining the optimal program x, the following vectors enter into the basis a 1 - 7 ' 1 ,..., aljo, a 2 j l , . . . , a 2 ^ , where a + /3 = m + l. According to the duality theory, an optimal program of the dual problem is derived from the system 7ri + ^a\n 7r a^h Tm irm+ +dYdiu cxn , , aa\inj , 7Ti 7T22 + ■ ' ■■• •+ +CkL UJ ==C^ 4j^i ai^ 7 1 "! + + a\]a*i a
L i W l + ahin2
+ ■■■ + amjaTrm +
h a
mn n*™ ™ ~-
+ d3au A A
w == °W
a^TTi + Oj^irs oj^irs + • ■ • + « i j > f f » -
A
= c)a , XX cC
° hh '
ow =
A
oC^ oc^
by Kramer's rule. Let us calculate the determinants: a
a
\n
\n
J 2
a
°
h? •*fr
••■
amjl
I" a
he 4»
dn = A = g(\0)
=A = 9(\o) == fcA kX00++1.l.
r ■•• ■••
a
mJp «*£
A
-Ao ~ °
Application
of Solution
Considering nonzero.
that
Techniques for Large Dimension
Problems to . . .
the problem is nondegenerate,
the determinant
For 1 < i < m a\nh
a\n
...
^
...
a
a
L
■••
C
••■
d
J«
== AAI t == /i/it (A t (A00))
a\n afj,
af a\nis
... ...
Aoc^ \QC\
... ...
-A - A 00
= Pi-^o Pi^l + qiX0 +U +9tAo +n
a
a
••■
A
•■• - A °
L
he
lw
L
°S>
dn
and for i = m + 1
a
i]a
a
2ja
■■■
a
c Ja
A
mJa
a?n
4fc
■•■
a
aa
aa
••• ■■•
il WW
L L
™h
Amm++1i =— //limm++ li ( A 0o)) -— A — i = 99m+iAo m + i A 0++ r mrm+ +1
°4 XX Cc
0° ii
and the dual estimates are equal Pi\l + + gjA qi\0 0 + + rin _ Pi^l u\ i ; ' ~ kX0 + l KA 0 4- ( 7m+iAo + rrm+ m+ii <7m+iAo + LO = . A:A k\ + 11 00 +
K Ki *~
. i = 1,... ,m ,
i =
l,..-,m,
T h e vector x also depends on A0 l x = A~ A^b b
Calculate A^1
to find out in what way x depends on Ao/
a\h
...
a\Ja
a\n
...
a
■ • • ■• ■
a a 2ja2ja
a a
• •■■ ■ ■ a 2 j , 3 a2j0
a amj1 mj1
■ ■ ■ ■ ■ ■
dnrfj,
■. . .• • dJad ^
2j 2 xJ !
2 j2h i
a\J0
>
A basic = = A basic \\
a amja mja
a amji
mji
-AQ -A 0
■ ■ •• ■ ■
. .■. . -
a "mjp mjp
- A0 0 -A
/j
279
is
280
Systems
Optimization
Methodology
det Ab = A = kX0 + I. The inverse matrix elements are calculated from the formula _ ffji(Ao) _ kjjXp + Ijj Ajj ~~ det Ab ~ kX0 +1 ~ k\ kX00 + l '
a3
'
where AJt is a cofactor to an o^ element of the initial matrix. Thus A^1: //knXo /cnA0 + + hi kX0 Q++1 l fcA
k2iXo + hi kXkXo 0 + l+ l
kl2Xo + l\2
kmXp + I22 hi ^22-^0
kXo + l klm
kmiXp + lmi "'"' kXo + l+ l kXp fcm2Ap fcm2Ao
kXo + l
kXp kX0 +1 +1
+ lm2
"'
+ l Ao XQ + + 'l l m ++ 11 fem fem + 1 A Xoo + + ljm+1
\
W n \ kXkX 0 +1 0 +1 L + H2 lm+1
kXo + l
fcmm
kX0 +1
+ lAp + /mm + 1
kXo kXo ++ll
kXp kXo++l
l
'<m m + 1l m + 1
kXp kXp+ +l
l//
and
E,
6
kijXo + hj
'
T
J€baS1S
fcA0+/
'
i=l i.e. S j are the linear fractional functions from A0. From the optimality criterion it follows that the vector x remains opti mal for all A, with which the following is accomplished:
2
2
1i Pi A2 -t-giA + + rt rt j paA ^ A 2 + qq22A A ++ r 22 aa nn + ++ + a 22 11 _
*A + J fcATJ
"
+ a
" ° 1 p m A 2 + gmA +
-
kxTi
1 PiA22+(?iA +giA + rx
a i n i1 _
"
1 PmA2z + — i fcA 2 2 P1A -HgiA + rr-! +g1X + 1
»" ~ ^fcATl *A+i 2
a
"
7 n
~fcA fcA+ i
" ++' ■' "'
+ a21 a21
kXTl feA+7
++
'""'
2
A + qmX + A2 + qmX +
+^ 1 A + r 1 2 PiA 2 +g
^
Cl ' ^
kx + i
r , <7m+lA i +1 j m+m gqmmX A+ + rm m 9 m + l A + rT" +dni n i +dni S c + / JkA + / ~ kX + l " > '' 2 p2A + g2A + r 2 2 P2A
+ a flmni
a
+ dl
' ' '" " x + ' m+1 _ \
j p22A A22 + gq2A X + r2
|
++fal 22nni i
' ia+1 fcA + /
kX + l fcTT7
kx + ii
'
|
2
++a a
^
p2A2 +g 2 A + r 2
kxVl kxVl
|
++
""
Application
of Solution
Techniques for Large Dimension
p m A 2 +<7 m A + ™> IxTi kxTi
A+ kTTi kxTi
2
+ a
A
Problems to . . .
-Ac"2'
—>0 E bi6; —— fcA+ +l Z ~ ~ kX kijX lij K-ijX -+"h (tjr
J £ basis
;
i7 == 1i
Definition of the 5 solution stability domain: (1) Let kX + I > 0 and k > 0 The domain of stability is then determined by the inequalities m
TTi
{k\ + l)Ai = \X22Y^ {kX YaaaPi aPi
+A
1=1 i=\ 7n 777
a 19i ^2 XX li9i +diq + r f i gm+ c\k m +i 1 --c}/c L L i i==l l
a ri+dirm+i + !Yl Z 0 l*l i r i + d l 7 ' m + 1 +
- cC _
°' il'' -- °'
7i = = 1l
771
777
2 al
x
= X Y^2a\ (/cA + Z)A ni = iniP' nipi
+
7i = = 1l
L L i i==l l
771
i=l 777
{kX + l)Ani+1
= \2
a pi y^agPi Yi n
-~1rn+i gm+l —-c\k cffc
. 7 = 1 777
777 T77
+ A ^ oa 2f ii gf it -- rJ mW+ i1 — cf I ++ ^ a 2 1 r I < 0 , .. t = l fel JJ i=l i=l 777
2 (fcA Z)A nni+n2 i n 2 p9ii ~ - qm+1 qm+1 ~ - Cn cl*2kk i+n2 = (fcA + + Z)A = A A2 ^YLOa**** . 7 = 1 777
+ A A +
777
9i X] aa i"2 9i
X] i"2 .
7= 1
Tm 1 c l ~ ~Tm ++ 1 ~~Cnn22l
. i=l
J
J
7 = 1
i=l
777
777
|
777
777
I
b k b j A Yi i = l * ij + Y^ 7=1 ^
or
+~^2^n 2ri<0, +Yazn 2ri-0'
i=l
- °>
A>_
fc
7=1
£j; A2 + Cj X + MjA Mj X + Nj < 0
j = = 1,..., n
281
282
Systems
Optimization
m m
m m
i»=i -l
i=l 1=1
Methodology ii
The system is consistent, since it is satisfied with A = Ao- Define
._ Xl
-Mj-^Mj-iCjNj
~
2Z~ 2£~ -M, +
A2 =
'
'
JMj-ZCjNj 2£~
•'
For those j , for which Cj > 0, we have Ai = max A, , c,>o 1l Ao = min A A,9 ..
We have obtained the range of changes for A € [Ai,A2]. If among the inequalities CjX2 + Mj\ + Nj < 0 there is no such value for which Cj > 0, then we assume Ai = —oo, — oo,
Application
of Solution
Techniques for Large Dimension
Problems to . . .
283
A 2 = ++O000 .
For those j , for which Cj < 0 we obtain: A3 = max A^ A, , , £,
If for a particular jo Aj° = XJ2°, then the curve also lies completely under the abscissa. In this case, as before, we assume that Af = - 0 0 , Af = +00. Af" We have obtained the range of changes for A (— 00, A4] and [A3,+00). For those j , for which C3 = 0, we calculate \i A
~
- ^ Mi Ms '
define A7 = max X\3, £j=0 JUj<0
284
Systems
Optimization
Methodology
Ag = min Xj , £,=0 Mj>0
and obtain the range of changes for A [A7, As]. Let us consider the second group of inequalities m
m
Xy^bjkjj h k^iJj >- 01° ' X ^ hhj ++^22 i=l
J e6 basis.
i=l
Let for J = 1,... } e
m
771
^2 Y^ hhj > 0 i=i
and for J = e + 1,... ,m+ 1 m
Y,bik*J
<0
-
7=1
m
(The case E kkij = 0 is not considered.) Calculate i=l
771 777
E M« hhj
\J — '=i — ~~m
A
E kkj E M*J
1=1 7=1
Then there should be A > XJ J = 1 , . . . , e and A < XJ J = e + l,... , m + l . Define A5 =
max ^
XJ,
&■*,./>o
1 = 11
A6 = Ag
min
AJ
f2b'k'J<° .=1
We have obtained the range of changes for A [A5, A6]. The intersection of all the ranges obtained gives the domain of solution stability.
Application
of Solution
Techniques for Large Dimension
Problems to . . .
285
Define A = max< A i ; A 5 ; A 7 ; - - > , A = min{A 2 ; A4; A6; A8} , A = msxlAr,A»;A5;AT;--> , A = min{A 2 ; A6;A8} . Then the x solution stability domain [[A,A] A, A ]UU [[ A A,JA (If A or A is achieved on —l/k, then the bound is not included in the domain, e.g. (A, A]. If, say, A < A, then [A, A] = A). Therefore the domain of stability of the solution x is convex when either all C3 > 0 or A3 = —oo and A4 = +00. (2) Let kX + I < 0 and k > 0. Then the domain of stability is determined by the inequalities CjX2 + -hMjX Mj\ + Nj Nj>0 > 0 m m m m A V* bikij btkij + J ^ btUj lij < 0 i=l
j3 & £ basis^ basis'! 1 J € basis |
i=l
>t
We determine from similar reasonings A'1, = min A, , X\ £>>o ' X'0 = max Ao , X\ = min X3,-. , , 6
Cj<0
l
X', = max Ai , X'5 -
min .= 1
XJ,
286
Systems Optimization
Ag6 = A
Methodology
A \ J7 ,
max
f^hku
Ao = max \A3 J . 88 £J=0 c,=o Mj>0
To obtain the solution stability domain, intersect the obtained line segments. Define A' = max{A maxfA^Ag;Ag} 4 ;A 6 ; A8} , A = min<j--;A'i ;1A3;A' ;A 3 ;A' 55;A7 A'=minj-^A' ;A;J> , max{A2;A4;A6;A8} , A' = max{A2;A!i;A
I = minl--;\'3;\'5;\'7
I
Then the domain of stability of the solution x [ A '' ,,A A' ']]UU[ [AA' ,' A J '' ]
(If A or A is achieved on —l/k then the bound here is not included in the domain, e.g. [A,A ). If, say A < A', then [A',A ] = A). §34.3. Solution Algorithm Step 1 :
Suppose we have the problem
{
cx
EdjXj
3" 1Z
Y. ) , + - k
n2
£^if -
J=I
A(jr,y) A(X,Y) jr > X > 0o,,
= = fc, b, F F > >o 0 ..
)
Application
of Solution Techniques for Large Dimension
Problems to . . .
287
Replace it by the problem min
\c22V) Y) (c*X + Ac
min dX-\ecY=0
xX >>o, 0, Step 2 :
y >o.
Y >0.
Solve the parametric problem l 2 2 min {c (c*JT X + + \c Ac Y)y) ,
AiX ^ x +A ^2Yy = = b, 6, dX - A dX \eYe F = 0 ,
xX >> 0, o,
yF >o. >0.
Step 2.1: Prescribe A = AoSuppose for A = Ao there is the basic initial program (a;7i(Ao),xj2(Ao),-- -,a;ja(Ao),2/ji(Ao),.. ■,xJa(X0),yJ1(X0),... ■,yjp{\o)) , y ^ ( A 0 ) ). . Calculate A^1 for the program and fill in the simplex table with the changing A
x1
...
x"i
XJ1(\)
«Jit(A)
...
4 1 S l (A)
Q
xJa(A)
i ^ W
...
^Qni(A)
...
4a*1+n»W
ej,
!/Ji(A)
x=il(A)
...
xJini(A)
x5ini+1(A)
...
x5ini+n2(A)
yJ0(X)
x^(A)
...
x^„t(A)
x^„l+1(A)
...
i^„1+n3(A)
...
Ani(A)
A„1+i(A)
N
CJ
XJ
Ji
cj,
Ja
4
/l
x"1+1
...
x"i+n2
w • - ■Wf.w
J0
c
2
Jf3
y(A)
1 At(A)
where p fc = l
*:=!
... j
A„1+n2(A)
288
Systems
Optimization
Methodology
X>(\) = A;1(\)A{,
j = i,...,m,
n
x ^(\) >+>(\) = A^(\)Ai, A^(X)Ai,
j = l,...,n2,
a
0
Aj(A) = j y j f e 4 f e i ( A ) + £c 2 J f c x 2 7 f c i (A). k=\
k=l
Solve the problem for A = A0 by a direct simplex method. If the problem for A = Ao is unsolvable, i.e. the linear form is not bounded, then we define the set of all A for which the problem cannot be solved from the relations Xjw(A)<0,
fc=l,...,al
* = !,...,/?/
(the vector j should enter into the basis). If we obtain the optimal program, then from the table of the relations obtained we determine the domain of stability of this program for all A, for which A j fWf - la g- +^ f *f i "l f' l ;s o. A xj{\)
= YJb> i=l
kX +
i
>°
basis. it M
i t
- / e basis.
Step 2.2: Suppose that A crosses one bound of the domain of stability. Then either one of xj becomes less than zero or one of Aj be comes larger then zero. If one of xj becomes less than zero, then the problem is further solved by the dual simplex method in one table. The vector J originates from the basis, the vector, on which max < *■ \ is achieved, enters into it. We determine the domain of those A for which the problem has no programs from the relations xjj
> 0.
- If one of the characteristic differences Aj becomes larger than zero, then the problem is further solved by the simplex method in one table. The vector j enters into the basis.
Application of Solution Techniques for Large Dimension Problems to . . .
Fit 0 = min
X 7
xjj>Q X j j
=
Xlr
Xkj
289
, the vector k leaves the basis. Transform
the simplex table. - If simultaneously one of x j becomes less than zero and one of Aj becomes larger than zero, then the problem is further solved by the dual simplex method so that it would always be in a feasible region. Transform the table. Step 2.3: The new problem obtained is tested for optimality. - If the program is optimal, then we determine the stability of the domain once again. Go to 2.2. - If the program is not optimal, the problem is further solved according to three possible cases till we get the optimal program. Determine its domain of stability. Go to step 2.2. The problem solution is completed if for each A from (-oo;+oo) we obtain the optimal program or see that the problem is unsolvable.
290
Systems Optimization Methodology
Step 2.4: T h e solution of the parametric problem results in the function of the objective function values depending on A, which looks as alX \22 + + biX bl\ + + Ci C1 ( ai »1 A + (1 d\ A + 1 \ 22 q 22 A + 62A + C +&2A C22
**W=< W=|
d2x + ll22
v
.
. ,.
A€(-oo,AJ A€(-oo,/?iJ .
,.
ay fli
** €€ [[ A A .. ««
a t A + tt
or
does
not exist, when for such A the problem has no programs at all. Step 3: We have mmip(X)
if among ?i(A) there are no such
that
ipi(X) = - 0 0 . (a) dt = 0 - If di/li > 0, then the function has the minimum at the point A = —bi/2di. If
A = -2a ;£-*[A-i,ft], t
then the function takes its minimal value at the ends of the segment line. - If di/U < 0, then the function assumes its minimal value at the ends of the segment line. - If Oj = 0, then it is equivalent to the case 6.1. (b) 4 ? 0 The case, when the point
-|-e[A-i,&], can not occur, since there would then be
^(-^) = -°°'
Application
of Solution
Techniques for Large Dimension
Problems to . ■ ■
291
which means that the linear form is not bounded on the set of programs, and this shows itself even in the process of solving the problem. Derive a,X2 + blX + c1 _ Oj ftj di\ + li diX * dtX di\ + lil% (1) Qr = 0 is a linear function, it assumes its minimal value at the extremities of the segment line. (2) 7i = 0 is a monotone function, it assumes its minimal value at the bound of the segment line. (3) 7 l ^ 0 , f i t ^ 0 . - If jiCli < 0, then the function is monotonic, it assumes its minimal value at the bound of the segment line. - If 7;fii > 0, then the function has the minimum at the point lidr
__ -Uli + Vjjftidj li
Step 3.3: The obtained A*o corresponds to the optimal program x(A*0). Terminate. The program x(A*o) is optimal for the original problem. R e m a r k : The more complex problems, where the parameter A enters into more than one constraint, are not solved by the present algorithm, since we would have to solve then the inequality of higher degrees.
292
Systems Optimization Methodology
§34.4. Block Scheme
.Application of Solution Techniques for Large Dimension Problems to . . .
293
294
Systems Optimization Methodology
Application of Solution Techniques for Large Dimension Problems to . . .
295
296
Systems Optimization Methodology
Application of Solution Techniques for Large Dimension Problems to
...
297
298
Systems Optimization Methodology
Application
of Solution
Techniques for Large Dimension
Problems to . . .
299
§35. A P P L I C A T I O N OF T H E ALGORITHM TO T H E G R A I N F A R M I N G OPTIMIZATION PROBLEM §35.1. Immediate application of the above parametric programming me thod to the problem of grain farming optimization is possible due to the nonlinear character of constraint 72. But the inequality can be replaced by two inequalities (which is more exact), where one of them stands for employment of the annual fund of the operating time for storage, i.e. 31
67 a
XX
&72 > 2 J T2i ?2J JJ + X ^ j=29
a
j=62
^3XJ ■
In the present equation, 0727 denotes the operating time expenditures for storing 1 ton of grain for 6 months, and £72 the annual operating time fund for grain storage. The second equation takes the form: 79
&73 > &73 > X ^
j = 73
aa
X
^3 ^3XJJ
1
where aj^j are the operating time expenditures for drying 1 ton of grain or for preserving 1 ton of grain for 1 day, and 673 is the daily operating time fund for drying and preservation. Evidently, it will be used at most in the last day of harvesting when the preservation capacity and the drying ca pacity are fully employed. The difficulty is encountered only in determining 672 and 673, since these magnitudes separately have not so far entered into statistics. The feasible region now is the convex polyhedron determined by 73 linear inequalities. With the objective function - maximum profit and the objective function - home production maximum, the problem was turned into a linear programming problem which can be solved by the simplex method. With the objective function - self-cost minimum, the problem was solved by the parametric programming method. §35.2. We consider a problem in the self-cost minimum.
300
Systems
Optimization
Methodology
Denote by 18
/ , djXj A A
~ ~
77
E X*j E J
jj = = 77 33
18
djXj E <%*j r(A) = r(X) = 345 + (1 - s ) ~ ^ = 345 + (1 - s)X ,
E E
j=73
xx
ii
where s = 0.1. 18
E djXj ^ j 3 = 27* + (1 - t ) ^ = 27t 27t + + (1 (1 -- t)X t)A ,, M (A) = M (A) = 27* + (1 - t) ~ E «>*j ■ j3==73 73
where t = 0.1. Then constraints 74 and 75 taken together with restrictions on 5 and t change to r(A) = min{34,A} min {34, A} , n(X) = /x(A) = min min{27,A} {27, A} Let us formulate the problem in terms of new designations
{
72
78
^
2 j c ^ X j + A ^ J (?jXj c^ij + + r(\)Cj r(X)cjg9Xx7g j=\ j=73 i = l J = 73 ^79a ^ X j > b i , i = 1,...,73 79
>, J
"Y^dijXj >bi, i = 1,...,73 i=i (the inequalities with the sign < are multiplied by —1) 18
J£ ] djXj djXj A= ^
r(A) r(X) = = min {34, A} , ,
x
E ^i E j=73
M(A) = min{27,A} , x,; x,; >> 00 ,,
jjf = = 1l , . . . ,t 7 9 .
Application
of Solution
Techniques
The relation
for Large Dimension
Problems
to . . .
301
18 18 ££
djXj djXj
tl £*
77 77
_\ -A
--
£ XJ
A A
j=73 j=73
can be varied according to its economic meaning only in the interval. 20,50 §35.3. We substitute the problem with an equivalent parametric program ming problem
{ 79
72
78
+ A
53 ^
j=l
3=1 ctijXj > 6^,
>79 i>= i
ctijXj
i=i
18
"I
Y l C)X1 + r(A)c?9*79 >J ,
j=73
i=73 i = 1,..., 73,
> 6^,
J
t = 1 , . . . 1731 77
218J d?1.? ~ A /77, Xj■ = 0, = 73 x = o, 23 ^^ - Aj y^ 7
j= l
j = lr(A)
j = 73 = min{34,A} , r(A) = p(A) = min{34,A} min{27,A} ,, p(A) = min{27,A} Xj>0, j = l , . . . , ,7 9 .
Xj>0,
j = l,...,79.
Considering that the parameters r and fi depend on A, the process of solving the internal parametric problem has to be subdivided into three parts. We solve first the problem
(
79
72
79
"I
2 C X 123 C ^ + A+ A Y^13 ) 4*3 i\ %**
2 j a^Xjf=l > 6i,
J , i J== 73 1 , . . . , 73
79
J=I
J ^ o y18 x ^ k , d A J=I
i = l,...,73, x
77
13 ^ - 53 > = °
18 /=!
53 «^*i
%>0, %/ =>! 0 , where A € [20,27].
77
_ A J = 73 X
13 3 = 0
j = l,.-.,79; j J=73 = l,...,79;
\
(35.1)
^35 ^
302
Systems
Optimization
Methodology
As soon as A crosses the upper bound and A < 34 is satisfied we change the objective function coefficients, i.e. turn to solving the problem
{ 79
y
72
79
~)
H C)X3 + X J2 4Xi \
jJ= l axjXj > 6,,
33 ==73 73 JJ i = 1 , . . . , 73,
(35.2)
79 3=1 / _, QijXj > 0; , I = 1, . . . , to , 18 77 3= 1 d x _A X =0 18 77 3 = 11 3j = 7 3
Ylr i i
/„c o)
Yl J
Xj>0,
j=l,...,79.
For this purpose, we calculate new characteristic differences of the optimal program i ( 2 7 ) of problem (35.1) from the formula aa
0
fc=l k=i
*w=l k=i
A](A) = ^ 4 i J l ) ( A ) + ^ 4 x 2 J t ) ( A ) , where cjfc(A) are the objective function coefficients of problem (35.2), and solve problem (35.2) until A crosses the upper bound of the interval. 2 7 , 3 4 As soon as we have A > 34, we similarly proceed to solving the parametric problem for A 6 [34,50]
{
72
78
1
C Z c xs ] 3i + ^c)xj{j ^E 79 c3 ^ J + A ^2 «?
3=1 V , a-ijXj >33=73 h= 73, 79 3/ = 1 &IJ Xj -^ 0{ , 3 = 1 18 18
2
J3 == l1 3=1
d
> 79) \I ,
i = 1,..., 73, 2 — 1,...,/o , 77 77
J
(35.3) / ^ ^ ^ *\
^ - A)3==J73 ] x, = 0 '0
Zj > 00 ,, *i >
3 = 73
= 1 1 ,, .. .. .. ,, 7 79 9 .. ij =
T h e regular solution of problems (35.1), (35.2), (35.3) results in the func tion of the objective function values
Application
of Solution
Techniques for Large Dimension
Problems to . . .
303
It should be noted here that when solving the above problems we do not obtain as Aj the expressions involving A with a degree larger than the second, though the parameter enters into the objective function with a larger number of variables than in the constraint case. The above problem of grain farm optimization was solved experimen tally for 18 2_j CLjXj 2_j CLjXj
A = ^ -
77
=50
E ®3
j=73
by the three objective functions. The constraints proved to be consistent. The obtained solutions showed that after careful treatment of matrix coeffi cients and constant terms the problem could give results well suited for the process of planning. As noted above, the coefficients a, a, a and the operat ing time expenditures for preserving and drying grain particularly require a more exact definition. The problem has the shortcomings of a more general character. Thus, all the coefficients represent average values over the whole country. The present problem is a static model for the period of one longrange plan. The capacities obtained due to capital outlays, immediately go into operation. This is the starting point for the next study which aims at constructing a dynamic model of grain farming optimization wherein the investment freezing and changes in the coefficients of the matrix A and in the right-hand sides for the long-range plan period are included.
Chapter 9 MAJOR PROBLEMS OF MULTIOBJECTIVE OPTIMIZATION
§36. FORMULATION OF THE MULTIOBJECTIVE OPTIMIZATION PROBLEM The multiobjective optimization problem is formulated as follows: Define an optimal choice of the element x° € X by a multiobjective indicator f ( x ) = {f(x)\y}, y € Y, where X, Y are the sets of admissible elements x € X, y 6 Y and / € F is an objective functional. The multiobjective optimization problem amounts to the extremal problem for a scalar indicator, if the set Y consists of a unique element y° and an optimal choice of the element x° 6 X, is derived from the condition for finding the extremum of f(x) — f(x\y°) by x e X. We call the multiobjective problem a vector optimization problem, if a set Y = I consists of indices 1 , 2 , . . . ,m (m < +00). For other cases with specified set y, the multiobjective optimization problem will be regarded conceptually. From the mathematical standpoint the above formulation of the multiobjective optimization problem is incorrect. Since attainment of an optimum of the multiobjective indicator f ( x ) with x € X is impossible in that if the element x° € X is selected from the optimum condition of any of the components f ( x \ y ° ) , for y° 6 Y of the multiobjective indicator f(x) = { f ( x \ y ) , y € y}, then no possibility is left for optimization (with the chosen x° € X) of the remaining components. Therefore, 304
Major Problems of Multiobjective
Optimization
305
formulation of the multiobjective indicator optimization problem has to be improved. It is necessary to formulate the rule defining the concept of an optimum by a multiobjective indicator. It is customary to call a rule of that kind the principle choice of a solution to the multiobjective optimization problem. Investigation of multiobjective optimization problems involves develop ing methods for specifying and incorporating determinative elements such as normalization, convolution, priority, etc., formulation of the axioms and properties satisfied by the principles of choice, consideration of concrete principles of choice and analysis of their properties, e.g. improvability, char acterization, and stability. And, finally, it involves statement and develop ment of methods for tackling multiobjective optimization problems. Let us briefly define the main characteristics in use, proceeding from the assumption that components of the multiobjective indicator / are preset in a positive ingredient (when selecting the elements I E I their increase is desirable). Simultaneously, suppose that there exist and are finite the utilized operations of finding min and max both immediately by a: G X, y EY and in all the others employed hereinafter. Normalization will be taken as a unique representation of F in F which transforms an objective functional / € F into another element of a space F. Among the major normalization procedures, we shall indicate the following:
- Natural normalization Ut(x\y) - Reduction to one dimension v[ip]
f(x\y)-mm f{x\y)-mm f(x\y) ,. , . x ; —,. . , ; max. f(x\y)min f{x\y) max.}{x\y)-mm f{x\y) ' v[f{x\y)\ vty]; v[f(x\y)] = v[p(y)] v[p(y)) xx v[ip])
- Transformation to measureless magnitudes
f(x\y)/u[f(x\y)\; f(x\y)fv[f(x\y)\ \
- Substitution of an ingredient
-f(x\y);l/f(x\y); - f(x\y); l/f(x\y);
- Normalization of comparison
f{x\y)I max }{x\y)\ f (x\y)/max f(x\y);
- Savage normalization - Averaging normalization
maxf{x\y)max f(x\y) -
f(x\y);
x£X
f(x\y) j j ^E f(x\y) f(x\y) /(*l»)■ • y€Y
306
Systems
Optimization
Methodology
Convolution of components of the multiobjective indicator / £ F will be termed the representation G e { F - t i ! 1 } which transforms a totality of components of the multiobjective indicator / corresponding to objective terms y £ Y into an objective scalar indicator G(f(x)) = G[{f(x\y)},y £ Y] £ Rl. When tackling practical problems of multiobjective optimization we employ linear convolutions as well as convolutions of maximization, minimization, product and various Cobb-Douglas functions:
G(f(x)) P(y)f(x\y), G(f(x)) = = ]TY,P(y)f(x\y), J/GK y€Y
G(f(x)) = mm[P(y)f(x\y) min[P(y)f(x\y)
+ q(y)},
G(f(x)) max[P(y)f(x\y) G(Hx)) = max[P{y)f{x\y)
+ q(y)] q(y)},1
y€Y y&y
y£Y
(/(«))=n y€Y y€Y
G(f(x))=
PM/M*) '
I J l[[P(y)f(x\y)}q(y). [P(y)Hx\y)]q(y).
yev
The problems of obtaining and substantiating a choice of convolutions are studied within the theory of utility. The study of the concept of priority in multiobjective optimization is based on comparing objective terms y £Y, components f(-\y) of the mul tiobjective indicator / £ F and values of f(x) € F of the multiobjective indicator / £ F on elements x £ X. In this case it should be borne in mind that so far there has been a lack of precise definitions of priority corre sponding to awareness of the essence of the multiobjective approach in the optimization theory based on a clear-cut differentiation of comparability by importance of objective terms, preferableness (significancy) of components /(■ | y) and efficiency of values of f(x) £ F with x £ X. Comparison of elements Z\, z2 of a set (space) z is regarded as fulfillment of some binary relation of order N in the sense that there exists one of the following conditions: zi is better than z2 by N(z\Nz2); z2 is better than z\ by N(z2Nzi); z^ is equivalent to z2 by N(zi — z2 = zxNz2 /\ z2NZ\)\ Z\ is not better than z2 by N(z\Nz%}; z2 is not better than zx by N(z2Nzi)\ z\ is not worse than z2 by N(ziNz2 V z\ **■ z2); z2 is not worse than Z\ by N(z2Nzi V z2 «- Z\)\ Z\ and z2 are incomparable by N(z\Nz2) A
Major Problems of Multiobjective
Optimization
307
(z2Nz\) A (z\ is 22). The binary relation of order J¥ is here determined by the following set of axioms: (1) zNz
VzVz£Z £ Z
(2)(z(21^22) (2) V (1)^ A ^ i ) 1Nz2)v(z2Nz
(reflexivity); \fzuz€2£Z Vei,«a Z
(asymmetry);
(3) {z1Nz2) A {z(22^2:3) =>z)^z ztNz lNz2)A{z2Nz 1Nz 3 z
(transitivity);
(4) zzix -— 2222 <^=> {z2Nzx) <^=> {z («tiV«j) A (^ATzj) 1Nz2) A
(equivalence);
The binary relation N satisfying axioms 1, 2, 3; 2, 3 and 1, 2, 3, 4 is called, respectively, preference (a nonstrict order), regulation (a strict order) and order (a perfect order) over the elements of the set z. Hence, priority can be defined as follows: The relation of priority over the set Y of objective terms is defined as a binary relation of order N(Y) for Y and makes it possible to compare objective terms yx, y2 € Y by importance according to fulfillment of the condition y\N(Y)y%, i.e. yi is more important than y2 -*=> yiN(Y)y2, y\ and y2 are equally important -s=> yi >« y2 etc. The priority relation on components /(• | y) of the multiobjective indi cator with y £ Y is defined as a binary relation of order Nj on the set {/(' |y)}i V € Y a n < i enables us to compare the components /(■ \yl) and /(■ I V2) w ith 2/i,3/2 £ Y by preference (significance) in compliance with ful fillment of the condition /(■ | y\)Nff(-1 y2), i.e. the y\th component /(• | yx) is preferable (more significant) as compared to y2 - the component /(■ | y2) if and only if this condition is met, etc. The priority relation on values of f(x) £ F with x £ X is defined as a binary relation of order N(f) on the set jF = {f(x) : x £ X} and enables us to compare values f(xx), f(x2) £ T by efficiency (a degree of increase in / ) , i.e. the value j{x\) is more efficient than f{x2)
308
Systems
Optimization
Methodology
them one should be careful because it is necessary to take account of the correlated statement of priority on the objective terms, components and values of the multiobjective indicator. The principle of choice will be identified as the formation of the relation of order N(f) on T with which normalization, convolutions and inclusion of the priority specification are chosen on the sets Y, {/(■ | y)}, y G Y and T. Formulation of the principle of selecting N(f) enables one to determine in what sense the solution of the problem for multiobjective optimization of the indicator f(x) on the set X is understood and define a set of opti mal elements. Notwithstanding that many principles of choice have been formulated, the problem of prescribing and taking account of priority on a set of objective terms and components of the multiobjective indicator, basically, has not yet been considered. Some principles of choice and their associated optimality conditions of x° G X are given below: - Dominant result principle max/(x°|y) max f(x°\y) > > max f(x\y) »6V
Vx G X ;
y&Y
- Tradeoff principle x x mama x p 3P(y)e ep--Yl °\y) * *E E p(v)f( \v); My) P ■■ Ep(y)f( (y)f( °\y)= = PtofWv); y€Y
y£Y
- Nash principle
/*(») - 0 ,
VT/GF,
Y[f{x°\y)>l[f(x\y), y€K y€Y
y6V y£Y
- Overall efficiency principle
£ / ( s O | 0 ) > Ey£V /(*|y),
yeV y€V
VxeX;
y£V
- Equality principle
/(*V) /(*V). VVy',y"eY; ye^i f(x°\y') = = f(x°\y"),
VxGX;
Major Problems of Multiobjective
Optimization
309
- Principle of Jeffry's efficiency inproper 3x€X:f(x\y)>f(x°\y) 3* € X : / ( % ) > ) f(m5\y),
Vj,eF,Vfc>0,
33f/ €£Y, F ,
3x € {x 3x {x G e X : f(x\y) / ( I | J , ) > / (f(x°\y")} *V)}
/f(x\y")-f(x°\y") W W W )
// W WW W) W W )
,,
'
'
w
,
c V
.«,I,#,^, * / „.0,I M 0 | ,
^€y:«*l»><^l»)5 ^€Y:/(«|y)(^|f)i
- Principle of Jeffry's efficiency proper 3z € X : / ( * | j ) > /(x°|y), Vy G 7 , 3x<EX:f(x\y)>f(x°\y), ¥y€F, 3 / - o - /(»■»)-/(»°l») << fc ft> 3fe>Q/(»■»)-/(»°l») fe >
/WWW)" /W) - /W) ~
VyGF, VT/GF,
Vi G {a; G X : / ( » | j ) > f(x°\y) 0
Va:e{a;eX:/(x|2/)>/(x |y)},
3y'€{y£Y:f(x\y)
My Vy G £Y F ;;
- Pareto's principle ATdom 0 33x ^G1X: : {{/(*!») /(I|J,) > > /(x°|j/) / ^ ° ( f ) V» V» G 6 rY}} A A {{ 33 // Ge Y Y :: f(x\y°) f(x\y°) > > f(x°\y f(x°\y0)} )} ,,
- Partial dominance principle 3Y f(x°\y) 3Yi1CY: C Y : f(x°\y)
> f(x\y) f(x\y) >
Vx G G X, X, Vx
y€ € Y* Y* ;; y
- Sure result principle minf(x°\y) min f(x°\y) y€Y y€Y
> minf(x\y) >y&Y rain f(x\y)
Vi ; Va; G GX A";
3/GV
- Least bias principle
ll/(*°W*ll < I I / ( * W 1 Vxex n/(»°)-riis]i/{*)-n vi€i r = {/*(»)}, {f*(y)},yeY; 2/ € F ;
/*() = max/(a%) max/W) igX
V
310
Systems
Optimization
Methodology
- A-criterion principle N\(f) x° G Xxo : X° A0 = = max (A|X (\\XX A # / A) 0
Xx = {x G X : A/(x|A) > A
VT/GF};
- Gurvitz a-criterion principle N"(f) ix°° 6GX X : : [ a min X/(*°|y) X/(s 0 |y) + (1 - Ot) a) max A , ^ 0 ^ ) ] A/(z|t/)l > [aminA/(i|2/) + (l - a) min A/(xb) L
yer
y€V
J
Vi G X X; Vx
- Uncertainty function maximum principle ix°° G X X : H[f(x # [ / ( z00)}) ] > > H[f(x)]
VxeX.
Define other concepts employed in solving multiobjective optimization problems. The element x° G X is called Pareto-optimal (unimprovable) by the multiobjective indicator / G F, if there is no J € X such that conditions fix\y) > fix°\y)> Vy € Y are met, if only for one y' G Y this inequality is strict. In such an event, x° G X is termed dominant (optimal according to the dominance principle) if inequalities f(x°\y) > f(x\y) Vx £ X, y £ Y are completed, i.e. the dominant element x° G X is Pareto-optimal. Define a natural nonstrict order Ndom on T as an intersection n Ny yer of the binary relations of orders Ny on {f{x\y) : a: G X } , y £ Y, i.e. f{x1)Ndomf(X2) with x , , i 2 e X « / f o l i ) > / f o b , ) , Vy G Y The natural nonstrict order JVdom is called nonstrict order A rdom , such that /(x 1 )iY" d o m /(x 2 ), xi, x2 £ X «=* /(ii)iV d o m /(a;2) and there exists if only one y' £ Y such that /(xi|j/') > /(x 2 |j/')- Since the relations of order are determined immediately on the values of the multiobjective indicator / € F, they (Ndom and 7Vdom) are called natural. These orders are decisive in the principles of choice in the form of dominance jV dom and Pareto-optimality (Ndom\Ndom) = Ndom A Ndom The element x° G X is termed improvable by the multiobjective indica tor / G F if there exists x G X such that inequalities f(x\y) > f(x°\y) Vy G Y are completed, and if only for one y' £ Y this inequality in strict, in other words, x° G X is improvable <^> 3x e X : f(x)Ndomf(x°).
Major Problems of Multiobjective
Optimization
311
Define the compromise region by / € F as follows
R(f) == \x°eX:J2 \x°eX:J2 R(f) I
p(y)f(Av) P(y)f(Av)
= max max J2 J2 P<3l)f(*\v) v{y)J{x\y) Vp(y) Vp(y) €€ 5511,, = X€
y€V
y€Y
)
where 5 denotes the space of pseudoprobabilistic distributions p = {p(y)\ y £ Y} on Y such that p(y) > 0, Vy e F , J3 p(?/) = 1 is completed. yev Definition of the tradeoff region i?(/) is conditioned by S. Karlin's re sults concerning establishment of relation between Pareto-optimality and linear convolution optimization in vector optimization problems with Y = I = { 1 , 2 , . . . , m } . Thus, if for some x° € X, p° = (p°,...,p°m) 6 5 the m
m
condition ^ pf/i(&f) = max J^ P^fi(x) is met, then x° is Pareto-optimal. I=1 If the set Xi=lis convex, the^sx functions fi(x) are concave and x° 6 X is Paretooptimal, then there exists p° = (p°,..., p^J € 5, p° > 0 (s = 1, m) for which m
the maximum condition holds for x £ X of the linear form ]T p®fi(x). i=i
Let us set forth the augmentation of Karlin's results obtained by V. V. Khomeniuk 7 for conditions of the conceptual formulation of the multiob jective optimization problem. T h e o r e m 3 6 . 1 : For x° € X to be Pareto-optimal, it is necessary that for any p(y) e S
Ylp(y)f(x°\y^l2p(y)f(x\y)
yey y€Y
yev yeY
v*ex.
T h e o r e m 36.2: For x° € X to be Pareto-optimal, it is necessary that there might exist p°{y) 6 5, for which T, P°(v)f(x°\y) ' ■
yeY
= max J2' P\vm*\v) xfc A T
■
yeY
T h e o r e m 36.3: For x°e X\[R{f)] to be Pareto-optimal on a set X\[R{f)\, where [R(f)\ is a closure of the set R{f), it is sufficient that for some
Ay) e 5 maxV £ pp°°(( yy )) / ( x | y ) = max £ p(s/)/(x|j/) p(y)/(x|y) . max V p(i/)65 x€X I6X xeXeX p{y)&S IL VTY * yTY £y VTY
312
Systems
Optimization
Methodology
Corollary 36.1: The set of Pareto-optimal elements x° € X on X belongs to the tradeoff region R(f). Besides, theorem (36.3) formulates a sufficient condition for the exis tence of the Pareto-optimal element x° € A\[i?(/)] outside the closure of the trade-off region. Theorem 36.4: For the set of Pareto-optimal elements x° £ X on X to coincide with the trade-off region R(f), it is necessary and sufficient that there would not exist p°(y) € 5, for which the following condition is met max
max x x £
Y\p°(y)f(x\y) y2p°(y)f(x\y)
~C
.. y ys sy y
= max J J
p(y)es
max V p(y)f(x\y) ■• maxY]p(2/)/(i|3/) L L
tcGX
"
J J
y€Y y€Y
Consider axiomatic problems of a rational choice. As far as multiobjective optimization is concerned, notwithstanding numerous works addressing the problem, so far there has been a lack of clear and sufficiently rigorous studies on the development of the axiomatic approach. To this point there have been formulated principles for selecting the elements / 6 F optimal by the multiobjective indicator x° 6 X, though it has been done, as a rule, without revealing the systems of axioms and properties that should be sat isfied by these principles which are fulfilled. Within the multiobjective optimization theory there arise the problems remaining most commonly unsolved: what are the conditions (axioms) to be satisfied by the principle of choice as a relation of order 2V(/}? Do the principles of choice satisfying these axioms exist? What are the principles of choice to be taken as a guideline within the accepted system of axioms? The basic systems of axioms have been stated in Arrow, 1,2 ArrowGurvitz, 3 V. V. Khomeniuk,7 Milnor,9 Nash, 10 as well as in Sen. 12,13 In this case it has been shown that these systems of axioms are satisfied accordingly by the following principles of choice in the form of specification of the relation of order N(f) : N(f) = A, N(f) = nyNy, N(f) = Uy[f(- | y)-/*(lf)]. N(f)
= minf(\y), yEY
N(f) = N({mmf(-\y), y£Y
max/(-|y)}), where f*(y) is y€Y
the status quo element of a set TF = {f{x) Vx 6 X : / € F } , usually defined as f*(y) = minf(x\y) Vj/ 6 Y and iV({min/(-1 j/),max/(- \y)}) dei£X
y€Y
yeY
notes the relation of order on the set R2 of the pairs of real numbers (fc, A')
Major Problems of Multiobjective
Optimization
313
satisfying the condition (ki > k2) A (Kx > K2) =>> (ki,K2)N({minf(-
\y),
y€Y
V. V. Khomeniuk has formulated in his work7 modification of the Arrow system of axioms satisfied by the principles of maximal efficiency by N\(f) and those of parametric maximal efficiency N%(f) with any p £ [0,1]. Arrow has formulated the following five natural axioms to be satisfied by order N(f) of the principle of choice. (1) Axiom of independence from positive linear transformation of the mul tiobjective indicator N(f)
^
N({[(m\y)f(-1 y) + *(»)]},» n(y)]},y G Y) Vm(y) Mm{y) > 0,n(y) G R1 iV({[(m|j/)/(-1y)
(2) Axiom of independence from the choice of elements x G X, i.e. if for two multiobjective indicators {f1(x\y)}, f1(x) = {p(x\y)},
yEY; = {f(*\y)}, {f(x\y)} i / G F ; f/ 2, ((x) *) =
,
y GF Y 2/€
elements x i , x2 £ X are such that there are
/Vh/) > //Vh/), V h / ) , /Via) /VbO >> f\Av) /Vis) Vvy y €€ Y, Y, and exist yu y2 G Y, for which Pix^y1) > / 1 (a: 2 |i/ 1 ), f2(xx\y2) > 2 2 1 1 2 1 P(x \y ) then we get x N(f )x , x N(p)x2. (3) Axiom of universality of N(f) by / G F, i.e. the principle of choice of N(f) must be defined for any multiobjective indicator / G F. (4) Axiom of the "dictator" absence: there is no y* G Y such that
1
2
^/)* ,i //(zV^/^V)^ ( • V J ^ / f a V ) ^ **#(/)** x 1 , x2 G X, i.e. there is no y" G F such that the condition x1N(f)x2 1 2 with x , a; G X is determined by the y*th component of the multiob jective indicator. (5) Axiom of the Pareto-weakened optimality: the condition jV dom = n Ny C N(f) is met, where Ny denotes the binary relation of y€Y
order in R1 established by the yth component of the multiobjective indicator / , i.e. if x° G X is optimal according to the dominance prin ciple f(x°\y) > }{x\y) Vx G X, y G Y then x° G X is optimal by N(f) : x°N(f)x, Vx G X.
314
Systems Optimization
Methodology
The well-known Arrow paradox resides in the fact that there is no or der N(f) satisfying these five axioms. Naturally it produces the desire to reformulate those five axioms so that there might exist the relation of order N(f) satisfying these new requirements. Thus, / 1 (x), P{x) are two multiobjective indicators of the element i 6 l , whereas N*, Ny are the two relations of order in R1 corresponding to the j/th component ^{x), f2(x) with y G Y, then the multiobjective indicator } 2 is termed the improvement of f1, and f1 the degradation of f2, if x1Nlx2, Vy G F => xlNlyx2, Vy G Y, i.e. if the condition }2{xl\y) > f2{x2\y), V y g F implies fulfillment of the inequalities fl{xl\y) > fl{x2\y), VyeY The changes in axioms 1-5 suggested by Sen consist in strengthening 1, weakening 4 and abandoning 5 which becomes here a consequence of axioms 2, 3, 6 and 7, and axioms 6 and 7 are formulated as follows. (6) Axiom of monotonocity (of positive correlation of the principle of choice): if, for elements xl, x2 G X the following conditions are fulfilled: (a) x1N{f1)x2 and f2 constitute improvement of f1, then xlN(f2)x2; (b) xlN{f)x2 and f2 constitute degradation of f\ then xlN(f2)x2 1 2 (7) Axiom of the free principle of choice: for any x ,x G X differing from one another there are the multiobjective indicators fl, f2 G F such that xlN{fx)x2, xlN{p)3? exist. In the above conditions it might be shown that axiom 5 is a consequence of axioms 2, 3, 6 and 7 and the system of Sen's axioms (axioms 2, 3, 6 and 7) holds for the principle of selecting N(f) if and only if it is "trivial", i.e. if it is the dominance principle N(f) — Ndom - fl Ny. y£Y
Thus, since the answer to the question about the presence of the prin ciple of choice (with the exception of the dominance in the class of order relations was basically negative, attempts were undertaken to develop new systems of axioms and substantiate, for them, the presence of the principles of choice in the class of binary relations of order. The system of axioms developed by Nash consists of axiom 1 (that of independence from positive linear transformation of / and /*) and the following three axioms. (8) Axiom of Pareto-optimality. (9) Axiom of independence from nonexistent elements, i.e. if X' C X x°N(f)x, V* G X, x° G X' then x°N(f)x Vi G X'.
Major Problems of Multiobjective
Optimization
315
(10) Axiom of symmetry N(f) <■* N(fi), where // is obtained from rear rangement of components according to the operation of representation t € {Y -+ F } . If the element /*(y) is status quo transformed into f*{y) in compliance with the operation I, then this system of axioms is satisfied by the principle of choice as N(f) = U[f(-\y) f*(y)\. y&y
Milnor has formulated a system of the following three axioms. (11) Axiom of incorporating the relation of order N(f) into ATdom, i.e. x°Ndomx*, x°, x* € X => x°N{f)x*. (12) Axiom of independence of fulfilling the condition x1N(f)x2 with x1, 2 x € X from the expansion X' D X. (13) Axiom of independence from the dominated elements: if for x' G X 3x2 £ X : f(x2\y) > f{xl\y) Vy £ Y, then the dominated element cannot belong to the set X° = {x° 6 X : x°N(f)x,Vx € X) of the elements optimal according to the principle N(f), i.e. the set X° does not depend on the removal of the dominated elements from X. The system of Milnor axioms is satisfied by the principle of select ing N(f) = min/(- \y), axiom 8 being, basically, the requirement for the dominated elements not to belong to the set X° of the elements optimal according to the principle of selecting N(f). Arrow and Gurvitz advanced the system composed of axioms 10 (of symmetry) and 8 and, moreover, including the following two axioms. (14) Axiom of independence from the removal of the recurring components f E F, i.e. N(f) — N(JM) where M is the compression of Y C Y into Y' CY such that Y' CY.y1^ y2, Vj/1, y2 e Y\ V 6 Y, 2 1 2 By eY':y =y (15) Axiom of independence from the removal of the inessential and some essential elements, in other words, if X\ C X2 C H = X, X\ nX$ ^ A then X° =Xxn X$, where X° = {a; {xQ0 £X X% = e X1x:x°N{f)x : « 0 JV(/)a: X\ 2 = {x° 6 X2 : x°N(f)x X°
WxEXx], VX GX:}, Vx 6 X2} .
The Arrow-Gurvitz system of axioms is satisfied by the principle of the guaranteed result, in which N(f) = minf(-\y), i.e. x°N(f)x, y€Y
316
Systems
Vi £ X <=> min f{x°\y)
Optimization
Methodology
= maxmin/(:r|y), a modified system of
y€Y
x€X y£Y
Arrow's axioms is composed of axioms 1-3 and the following axioms. (16) Axiom of the "dictator" acceptability, i.e. there can exist y* € Y such that fulfillment of the condition x1N(f)x2 with x1, x2 6 X is determined by fulfillment of the inequality /(a:112/*) > f(x2\y*) for the y*th component of the multiobjective indicator / . (17) Axiom of unimprovability of the elements optimal by N(f) i.e. if for x* e X the condition x*N(f)x Va; 6 X is fulfilled, then x* is unim provable (Pareto-optimal) on the set X\[X°] where [X°] is a closure of the set X° = {x° € X : x°N(f)x Vx £ X}. The above system of axioms is satisfied by the maximal efficiency princi ple in the form of N^f) and the parametric principle of maximal efficiency in the form of N?(f) with p € [0,1], and N?(f) satisfies the Arrow-Gurvitz system of axioms. Although the principles of choice proceed from the employment of convo lutions of the multiobjective indicator components, the formulated systems of axioms and the appropriate principles of choice do not touch upon the problems of definition, specification and incorporation of priority. Thus, in practical problems of multiobjective optimization one should be guided by the principle of maximal efficiency N^.(f) and the parametrical principle of maximal efficiency N?(f). Let us consider comparison of orders for various principles of choice. The binary relation of order TV on the elements z € Z of an arbitrary set will be referred to as extential if there exist elements z\, z2 € Z, z\ ^ z2 such that ZiNz2. If the relation of order N is extential, then it will be denoted by Nexi The relation of order N£xt is stronger than the relation of order N2xt, i.e. N2xt C Affxt if the intersection of sets is nonempty zi G Z :; z°N? z°NfAxtzz Zl = {z° £Z z2 = {z° {z° 6 Z :
z°/V2extz
Vz e Z} , Vz < eE Z} Z}
and if zi C z 2 , ie. there exists AT1extt>iV2sxt <=^ 7Vfxt C 7Vfxt <^=> {{zlC\z2 ^ A)A(z,C:2)}. It will be stated that the relation of order Nexi on z is extraextential if it is extential and the set z° = {z° € Z : z°Nextz Vz € Z} / A of the elements z° € Z being the best by Nexi is nonempty.
Major Problems of Multiobjective
Optimization
317
Since later we shall study only extraextential relations of orders the symbol ext for 7Vext is omitted, hence N e x t = N. Thus, the relation of order N is stronger than each Nm, m £ R if {N>Nm} and {z°r\zm} £ A, i.e. z° C zm where we denote via z and zm the sets of the form z0 = {z° £Z EZ : z°Nz
VzEZ}, Vz£Z},
zm = {z° eZ £Z : z°Nmz
Vz e€ Z} Z)
It will be stated that the relation of order iV is stronger than a totality {Nm}, m £ R, i.e. N > {Nm}, m £ R, if {N > Nm} Vm £ R and z° n { n zm} j4 A, z° C n zm. The totality of relations of orders {Nm\, m £ R is stronger than N, i.e. {A m } > N, m £ R if {A m > A} Vm G R and { n z m } n z ° # A, i.e. if { n zm—}Cz°. m€ff
meft
The relation of order A d o m is extential for the prescribed X,Y, f £ F, i.e. there are dominant and dominated elements in X, but the problem of choice according to the multiobjective indicator arises due to the fact that the set X° = {x° £ X : x0Ndornx \fx£X} of the best by A d o m elements is generally empty. However, the relation of order ^v d o m ^A d o m close to A rdom is usually extraextential, since it is assumed that the mul tiobjective optimization problem is nondegenerate, in other words, the set of Pareto-optimal elements a;0 € A" is nonempty, and the Pareto-optimal elements x° £ X are exactly the best by _/v dom \A rdom , because the prob lem of choosing the element optimal by the multiobjective indicator arises due to the fact that this set is sufficiently representative and, sometimes, coincides with X. One of the basic requirements for the principle of choice in the form of the relation of order N(f) must reside in the fact that N(f) would be extraextential. Moreover, it is natural to conjecture that N(f) is stronger than 7V dom \A rdom in the sense of the given definition, specifically, N„(f) > (Ndom\Ndom), NP(f) > (Ndom\Ndom) Vp £ [0,1]. The above concept > introduced on the relations of order AT enables one to undertake comparison of the principles of choosing N(f) as to whether there exist, for a specified principle N(f), other principles of choice in which their corresponding relations of order are stronger than N(f) or, conversely, N(f) is stronger than they are. Furthermore, for the concrete specified pairs of the principles of choice (or sets) it might be pointed out which of them is stronger and it can be done for the purpose of finding the strongest relations
318
Systems
Optimization
Methodology
of order N(f) with the prescribed X, Y, / . According to the principle of choice as N(f) we find for them the sets of the best elements which are assumed to be the solution of the multiobjective optimizatipn problem. The prescription of the principle of choice in the form of the extra extential relation of order N(f) for particular X, Y, f is specified by the system of axioms (conditions) which should be satisfied by the principle and, thus, it determines the system of axioms which should be satisfied by the principle. Hence we have two problems: the primal, which is to determine for the specified relations of order Nm(f) with m G R, the system of axioms to be satisfied by N(f)>{Nm(f)}, m G R, and how this system of axioms shall agree with the system of axioms for {Nm(f)}, m £ R, and the inverse, which is to determine for the given {Nm(f)}, m G R, the system to be satisfied by N(f) : {Nrn(f)}m&R t> N(f) and how this system shall agree with the systems of axioms for {Nm(f)}, m G R. So far the relation of order, N(f), for different principles of choice has been determined by choosing a convolution Cf(x) = {cf(x\y)}, y &Y, with x G X from the space of mapping the set X into Rl, i.e. cj G {X —► R1}, where c/(x) is a value of the functional c/ on the element x G X. In this case, for the prescribed convolution, c/, the relation of order Nc(f) = N(c/) is determined by a condition x1Nc(f)x2\x1, x2 G X <=>> ^ ( x 1 ) > c/(x 2 ). Hence prescribing a totality of the relations of order (principles of choice) {Nm(f)}, m G R is equivalent to prescribing a totality of functionals {cm}, m G R and permits the definition of a linear shell £o a n d a linear space C of the relations of order Nf{f) with p = {pm}, m G R as follows: £ 0 = { X p ( / ) } . V G S(R), C = {N?(f)}, p G S(R) where S(R) is a space of pseudoprobabilistic distributions on R S(R) S{R) = {p={Pm},m&R:Pm>0
S(R) = lp&S(R): II
J>
meR rngft
>0 m
Vm£R},
= ll J)
The relation of order 7VCP(/) with p G S(R) is derived by cj = £ in the form of N?(f)
= £ meR
pmNm{f),
thus: xlN^{f)x2\x\
m€R
x2 G X
pmcj ^
cp}{xl)>cpf{x2). Thus, by the systems of axioms Am for the principles of choice Nm(f) with m 6 R, it is possible to define the systems of axioms AP for the
Major Problems of Multiobjective
Optimization
319
principles of choice in the form of Nf(f) with p £ S{R) as well as the more general systems A(C0) and A(C) as a totality of the systems of axioms {A?}, p £ S{R) and {AP}, p £ S(R) respectively. It is the choice p £ S{R) or p £ S(R)\S(R) that forms concrete relations of order N£(f) £ £o and N£(f) € £\£o- Using the above definition of > as the base, it permits revealing whether with a specified p £ S(R) or p £ S(R) the relation of order in the principle of choice N?{f) is stronger than {Nm(f)}, m £ R and for which p £ S(R) or p £ S(R) the corresponding JVf (/) > {Nm(f)}, m £ R. Thus, following this way we can seek the solution of the stated primal and reverse problems. It is also possible to suggest other approaches to investigation of the above problems. §37. T H E S T U D Y OF T H E M A X I M A L EFFICIENCY PRINCIPLE Proceeding from the conceptual formulation of the multiobjective opti mization problem, suppose that for any y £Y the values f°(y) =min f{x\y), f*{y) = maxf(x\y) are finite. Employ natural normalization to elimi nate different scales for measuring the components of the multiobjective indicator f{x) = {f(x\y)}, y £ Y, thus we have pj(x\y) = [f(x\y) — f°(y)}/[f*(y) ~ f°{y)} where the inequalities 0 < pf(x\y) < 1, Vx £ X, y £Y exist. The value pf(x\y) with the fixed y £Y represents the degree of attaining an optimum, specifically, a maximum with positive ingredients of / at the point x £ X by the t/th component of the multiobjective indicator / . Let Xfj. be a set of elements of the form X^ = {x : X : pf(x\y) > pWy £ Y}, where /i is a prescribed value, 0 < p. < 1. In this case, the choice of a particular level p provides for the nonempty set X^ the degree of attaining an optimum by any one of the multiobjective indicator components being not less than p. It might be natural to define the maximal efficiency problem as the problem of obtaining a maximally possible level p°, i.e. we have to derive the solution of the ^-problem of the form p° = max {p\X^ ^ A). 0
Identify the element x° £ X as optimal according to the maximal effi ciency principle, if x° € XMo, i.e. the conditions p;{x°\y) > p°,Vy £Y are fulfilled. The value of p° is simultaneously the best possible approxima-
320
Systems
Optimization
Methodology
tion to an optimum by all components of the multiobjective indicator (the priority on components / is not specified). In vector optimization, one of the well-known principles of choice is the principle of obtaining a guaranteed result or the maximin principle the essence of which, as far as the formulation closest to the maximal efficiency principle is concerned, resides in deriving the optimal element x° £ X from the condition mmnfn(x°\y) vain = max| max fminnf(x\y)\ rnin pf (x\y) 1. . f(x°\y) = y€Y
xeX
I yEY L yEY
JJ
Thus, we have the following statements. Theorem 37.1: The maximal efficiency principle coincides with the max imin principle, thus /i° = max min/i/(x|2/) maxfmin/x / (a:|y)l.. xEX l y£Y
J
Lemma 37.1: We have the inequality M° > max[minp./(:r|2/) max|min/i/(i|y) . x€X
I y£Y
J
x € A L y£Y
J
Lemma 37.2: The inequality is fulfilled as max[min/i / (x|i/) > n°.
I S A L y£Y J maat|min/*/(ar|y)| > n°. i € A L y£Y
J
The maximal efficiency principle may be considered, on one hand, as a particular interpretation of the guaranteed result principle (maximin), and moreover as an independent principle of choice, being more "handy" as compared to maximin, wherein the game problem of optimization is stated and solved, whereas the /t-problem is the extremal problem for optimization of deriving /j?, x° £ X from the condition 0 /j, >p /j,0 = = max max {(j,\n {(j.\nff(x\y) (x\y) >/*
0<M<1 0<M<1
Vy Vy 6 eY} Y} .
xex xex
Test fulfillment of the axioms of choice for the maximal efficiency prin ciple. Thus, we have Lemma 37.3: The degree of attaining an optimum fif(x\y) at the point x e X by the yth component of the multiobjective indicator is independent of the positive linear transformation for any y £Y, »[m(y)f+n{ (y)f+n{y)](x\y) Vf(x\y) Vm(y) > 0, y)](x\y) = Hf{?\y)
n(y) £e R1 .
Major Problems of Multiobjective
Optimization
321
Theorem 37.6: The maximal efficiency principle satisfies the axiom 1 of independence of the principle of choice from the positive linear transforma tion of the multiobjective indicator f £ F. Lemma 37.4: The degree of success in attaining an optimum of fif(x\y) is independent of choosing the elements x £ X, i.e. if for two multiobjective indicators f1,f2£F the elements x1, x2 £ X are such that for any y £ Y the inequalities f^x^y) > fl(x2\y); }2{xl\y) > f2{x2\y) are fulfilled and 1 2 there exist y , y £ Y, for which the inequalities are strict, then for all yeY
vpi^ly)
1 1 2 2 vp{x2\y) <=*> ftpi® ]®)> >np{x np{x > »p(x <=>ppi® ]®) \y).\y).
T h e o r e m 37.7: The relation of order N^(f) for the maximal efficiency principle satisfies axiom 2 of independence of the principle of choice from the choice of elements x £ X, i.e. if for two multiobjective indicators f1, f2 £ F and the elements x1, x2 £ X the inequalities fl{xl\y) > /1(x2|y), 2 2 2 l / (ff%) > f (x \y) Vy 6 Y are fulfilled, then there is x N^{fl)x2 «=>■
xlN^f2)x2.
Theorem 37.8: The relation of order iVM(/) for the maximal efficiency principle satisfies axiom 3 of universality of the principle of choice by / £ F. Theorem 37.9: The relation of order N^lf) for the maximal efficiency principle may not satisfy axiom 4 of the "dictator" absence for the principle of choice, i.e. there are such multiobjective indicators f* £ F that y* £ Y is defined for which the fulfillment o(x1Ntl(f*)x2 is determined by the y*th component of the multiobjective indicator /*:
r(*V)>r(*V) x\x22ex. r(*V) > A*V) =*x'N^nx =*z1 W V2, x\x £X. Theorem 37.10: The relation of order N^(f) for the maximal efficiency principle satisfies axiom 5 of the weakened Pareto-optimality principle of choice. Theorem 37.11: The relation of order Nfi(f) for the maximal efficiency principle satisfies axiom 7 of the principle of free choice.
322
Systems
Optimization
Methodology
Theorem 37.12: The relation of order N^f) for the maximal efficiency principle satisfies axiom 16 of the "dictator" acceptability for the principle of choice. Lemma 37.5: The element x° € X is improvable according to the multiobjective indicator / € F (there exists x* £ X such that f(x*\y) > f(x°\y) V j £ F if only one of the inequalities is strict) if and only if there ex ists A°(j/) £ {Y —► R1} and the inequalities nj(x°\y) < a\o(y),^j(x°\y) < a\o\y*) are fulfilled for all y £ Y and if only for one y* £ Y, where a\°(y) = a>° ~ A°(y), a^o = a\o = maxmin[/z/(a;|j/) + A ($?)]. (y)].
Theorem 37.13: The maximal efficiency principle satisfies axiom 17 of unimprovability (Pareto-optimality) of the element x° £ X optimal for the relation of order iVM(/), i.e. if a;0 € X satisfies the conditions x°Nli(f)x Vx £ X •<=>■ min/i/(x°|y) = maxmin/t/(x|i/) then there is no x* £ JfVJf0 such y£Y
x£X
y€Y
that /(x*|j/) > f(x°\y) Wy £Y and if only one of the inequalities is strict. The proofs for the above statements may be found in the author's work.8 The above statements enable us to establish that the relation of order Nfi(f) for the maximal efficiency principle satisfies axioms 1-3, and axiom 4 may not be fulfilled for some multiobjective indicators / € F, The presence of Arrow's paradox and the result that with the strengthening of 6 of axiom 1 and the weakening of 7 of axiom 4 only the dominance principle satisfies axioms 2, 3, 5-7, and also the fulfillment of axioms 1-3, 5 for the maximal efficiency principle shows that when formulating a consistent system of axioms on a basis of 1-3, 5 one should modify axiom 4 by axiom 7 without strengthening axiom 1 and rejecting axiom 5. In this version the system of axioms 1-3, 5, 7 is satisfied by the relation of order 2VM(/) for the maximal efficiency principle. It is to be noted that axiom 7 apparently is "excessive" in the sense of weakening axiom 4, and the assumption of axiom 7 leaves open the question of whether to give up completely the possibility of a "dictator" (to demand its absence as in axiom 4) or to assume its possibility in some cases. The answer to this question is partly contained in the statement of theorem (37.12).
323
M a j o r Problems of Multiobjective O p t i m i z a t i o n
Consider the properties of the maximal efficiency principle.
(A) Linearity of the principle of choice: the relation of order N ( f ) may be presented as a linear combination of the multiobjective indicator components. The condition 3p(y) E P : N ( f ) = C p(y)f ( 1 y) is U - EY -
valid, and here P denotes a set P = { p ( y ) E {Y + R 1 ) : p(y) 2 p(y) = 1). If Y is connected, then the sum by y E Y is taken as 0, yEY
an integral.
(B) The membership of the region of the optimal compromise according to the principle of choosing an element. Thus we have x0N ( f )x Vx E X x0 E R, where the compromise region R is defined as follows:
*
R = {x* E X : 3 p r ( y )E P ,
(C) Weak dominance of the optimal elements x0 E X : xON ( f ) x Vx E X according t o the principle of choice, even though there is another principle of choice N * ( f )such that for a particular x* E X , x * N * ( f ) xV x E X . Here the conditions x O N (f )x' , x* N* ( f )xOare valid. We have the following statements.
Theorem 37.14: The maximal efficiency principle satisfies the properties A, B, C, the optimal elements x0 E X,o being dominant for N,( f ) with respect to the optimal elements x' E X according t o any principle of choice N ( f ) as distinct from N , ( f ) where
N P ( f )=
2;ynf (. I Y )
Lemma 37.6: If the components f ( X I y) of the multiobjective indicator are concave by x E X with y E Y, then
324
Systems
Optimization
Methodology
i.e. there exists a saddle point (x°,y°) £ X x Y for (j,f(x\y), x° being an optimal element according to the maximal efficiency principle. Using the maximal efficiency principle as the base, consider the charac terization issues of the multiobjective optimization problem. The characterization will be referred to as the parametric partition XT, YT with T £ T C R1 of the sets X, Y in the optimization problem of the multiobjective indicator f(x) = {f{x\y)}, y 6 Y when choosing the elements x £ X so that with an increase of T £ T there might occur a monotone increase (or, conversely, a decrease) in the fulfillment of a par ticular "property" of the elements x £ X and multiobjective indicators / T (x) = {f(x\y)}, y £ YT with x £ XT, or XT and fT{x) exhibit some "properties" differing from the properties XT>, fT>(x) with r / r', T now being unnecessarily from Rl. The form of Bayesian sets represents one of the very first characteri zations which has, comparatively speaking, long been utilized for decision making under uncertainty. In the adopted designations the form of Bayesian sets appears as T = X, r = x, XT = {x}, x £ X, YT = YT(x) = {y' £ Y : f{x\y') > max f{x\y)}, fT{x) = {f{x\y)}, y £ YT(x), where YT{x) is a set of dominant objective terms at the point r = x and fT{x) is a value of the multiobjective indicator f(x) after excluding the dominated components at the point r = x £ X. Based on the maximal efficiency principle, characterization of the mul tiobjective optimization problem is accomplished in the following way. Let T = 0, M r = 0 = 0, XT=0 = A, y r = 0 = A, XT=0 = X, YT=0 = Y, fT=0 = f. (1) obtain n^ from the solution condition for the /^-problem in the form fif - max \n\nf(x\y) 0<£i
> M Vy £ Y) = maxf min^ / (x|y)l , x€X ly€Y
suppose r = 1, / J T = 1 = $, thus we have X = Xi = {x £ X ; fif(x\y) > i4 Vy£Y], XT=l VT/ £ Y} , T=l = Xi = {x £ X ; iis{x\y) > M? Y =Y1={y£Y:»f(x\y) = YT=l = n° n°11 ^^ £ £ X,} X,} ,, T=l=Y1={y£Y:»f(x\y) fr=i(x) x£X; fr=i(x) = = h(x) h(x) = = {f(x\y)} {f(x\y)} ,, y£Y y£Ylt x£X;
J
Xj = = X\Xi X\Xi X x Y Y1 = =y y \\ ff tt 1
,, ,,
lt
(2) obtain jt° from the condition l4 = max [n\nfl(x\y) 0
>n
Vy £ ?i] = maxfmin/z/, (x|j/)l (x\y)] , igx Lygy
J
Major Problems of Multiobjective
suppose
T
— 2,
HT=2
Optimization
325
— M2> ^ u s we have
X £ X ::***(*]») iifl(x\y) > $ V j/efi}, =X1\X22, T=2 = X 2 = {x *,»a *a {*€* >*4 Vj/efi}, xX22=x,\x YT=2 Y {yeY:fi ==\°\° Vxe eXx2 }2 }, , ? y2 2==F fi\y i \ r 2 2, , 2 2=={yeY:fi h(x\y) 2 2 Vx T=2==YY h(x\y) //r= r = 22((i) i ) = /£(*) /(%)} , - {{/(^|2/)}, 2 (x) =
yye? € 2y,2 ,
x xeX. eX.
M2 > Mi M? > Mo = 0 is valid. We continue this procedure in a similar way. Suppose that for some positive whole number r € R1 there have been derived fi°,XT,YT,XT = XT^\XT, YT = f r _ i \ y T , / T (-) = {/(• | y)}, y £ y T , then obtain / i ° + 1 from the condition M°+i = max {/x|/i {fi\Hf/T (x|t/) > M // Vy € Y FTT}} = = max[ max [ min fif M/r(x\y)\ T(x\y) T(^|j/)1 ,. ix€X €X MT+1
-= M T + 1
assume that XT+1 = {x G X : nfr(x\y)
> n°T+1
WyeYT},
ryrr++11 = {2/ eyTT:M/.(^l3/) :/x/.(^l3/) = M°++1i Vxex T + 1 }, 2/ey / TT++ i(x) l(x) == {/(x|j/)}, Jj e 6 f r ++1i , x xex, ex,
XT+1 = XT\XT+1
,
yT+1 = y TT\y \r T+1 ,
where / i ° + 1 > ^ > M?-i > ' ■' > M° > A > Mo = ° e t c As a result, with T € T = { 0 , 1 , 2 , . . . , } c R1 we can obtain char acterization of the optimization problem of the multiobjective indicator / ( i ) = {f(x\y)}, y € y by x € X in the form of a sequence of charac teristic numbers fi® of the sets XT, XT of elements from X, the sets y T , YT of objective terms from Y, the multiobjective indicators / T € F . Thus we have T h e o r e m 37.15: The suggested procedure of characterizing the multiob jective optimization problem satisfies the following conditions: (a) monotonocity of the characteristic numbers n° with r € T increasing (decreasing), i.e. M T - I < M? < MT+I'-
(b) monotonocity of the characteristic sets YT of objective terms from Y, i.e. y T + 1 C y r C y r _ i with r € T;
326
Systems
Optimization
Methodology
(c) monotonocity of the characteristic multiobjective indicators fT € F with T G T, i.e. /r(-) »)} , / T (-) = {/(■ I yV)} ) } ,,» ! / e6 y?TT = {{/r-i(/ T _ i ( - ))}\{/(} \ { / ( - 1| y)},
2/yeY € FTT
therefore, / T is obtained from / r _ i by excluding the components y£YT; (d) monotonocity of the characteristic sets XT of elements from X, i.e. we are ''guaranteed" that the lowest degree of attaining the optimum of pPT on the elements x £ XT increases (decreases) as r € T increases (decreases). The proof of the theorem follows from the definition of the concepts utilized in the characterization procedure. The characterization process in the case of the problem of vector op timization of Y = I = { 1 , 2 , . . . , rn} (the characterization technique based on the maximal efficiency principle), here the set T contains not more than (m + 1) of the element and has a finite number of stages. The studies of the natural conditions of the choice in the form of axioms 1-5 and their modifications show that relaxation of axiom 4 down to the possibilities of the "dictator" presence in the principle of choice results in solving Arrow's paradox, as a theorem of impossibility, using the proof of validity of these axioms for the maximal efficiency principle as the base. §38. T H E S T U D Y O F T H E P A R A M E T R I C P R I N C I P L E O F M A X I M A L E F F I C I E N C Y IN M U L T I O B J E C T I V E OPTIMIZATION In a sense, the maximal efficiency principle is "conservative": even a drastic increase in the majority of the multiobjective indicator components may not involve an increase of the degree, optimal by this principle, of /i° = max min fif(x\y) of achieving an optimum. In this respect, the critique of the guaranteed result principle in the theory of decision making under uncertainty gave rise to formulation of the Gurvitz parametric criterion. As far as the problem of optimization by the multiobjective indicator is concerned the parametric principle of maximal efficiency is formulated as follows: for a specified p € [ 0,1] we need to obtain x° € X from a condition
Major Problems of Multiobjective
Mp = pmm fif{xl\y)
Optimization
327
+ (1 - p) max. fj,f(x°\y) = max \pmmp,f{x\y)
+ (1 -
p) m&x{j,f(x\y) , where p is a specified level of "weighing" the minimal and maximal possible degrees of achieving an optimum by the multiobjective indicator components. In terms of the principle of choice the parametric principle of maxi mal average efficiency leads to the following relation of order: N?(f) = pminfif(- \y) + (1 — p)maxp,f(- \y) moreover, for the optimal x° £ X we have x°pNP(f)x Vx £ X. With p = 1 we have the maximal efficiency principle, whereas with p = 0 we have the dominant efficiency principle for which a set of optimal elements is defined, respectively, in the form X J U = {x°
e X
'■ mm^/(ar°|3/) mm^/(i°|3/) = max mm mmpfif{x\y) f{x\y)
J,
X ° 0 = •! I x° x° ££ XX ::maxpf(x°\y) maxp/(i°|t/) — —max max max/x/(x|j/) max p,f(x\y) > \ I y€Y xex lyeY i) (
y€V
xeX
I y£Y
1)
For the dominant efficiency principle a set of optimal elements of Xp=0 coincides with a set of optimal elements for a dominant component of p.; (or / ) of the multiobjective indicator, thus we have X°p=0 = {x°eX:
0 3y° £ Y\ MHf(x°\y = maxpff(x\y°) / ( * V )) =
s€ A
= 1,
M /* ( *V ° |)/ ) > * / ( ^ W Vy£F} Vy£F} M >/M*°M Investigation into concrete problems of multiobjective optimization, as a rule, starts with finding a region X°(y) = {x° £ X : f(x°\y) > f(x\y) Vx £ X} based on the solution of extremal problems for each of the compo nents f(x\y) of the multiobjective indicator f(x) corresponding to the objective term y £ Y In most cases the multiobjective indicator com ponents are "inconsistent" (e.g. the minimal values of some components are achieved with the maximal values of the others), hence, it is natural that investigations of multiobjective optimization problems do not termi nate in deriving X°(y), y € Y. The values of the expression fij(x) = pm'mpf(x\y) + (1 - p) max p.; (x\y) for the fixed x £ X with p £ [0,1] yeY
y&Y
represent the interval Jf(x) = [p1f(x),p°f(x)] of the mathematical expec tation values of ppf{x) of the form (ipf(x) = pp){x) + (1 - p)p°f(x) for the chosen level of "weighing" p £ [0,1].
328
Systems
Optimization
Methodology
Now we consider a relation of the maximal efficiency parametric princi ple to the axioms of choice. Thus we have Theorem 38.1: The principle of choice N£(f) for the fixed p € [0,1] satisfies axioms 1-3, 16. The proof of the theorem is carried out similarly to that of fulfilling axioms 1-3, 16 for JVj(/) = N^f). Instead of axiom 5 we formulate the following axiom 18. The axiom of linearity of the principle of choice, i.e. the relation of order may be presented asN(/)= £p(y)/(-|y). y€Y
11 p(y)eS=\p(y)e{Y^R }:p(y)>0 P(y)eS=\P(y)e{Y^R }:p(y)>0 I
Vy€Y,
£>(») = l|. y€Y )
Theorem 38.2: The principle of choice N*(f) for the fixed p € [0,1] satisfies axiom 18. Theorem 38.3: The optimal elements x°v £ X for the principle of choice N?(f) belong to the tradeoff domain
R = I ix £ X : 3p(y) £ S, I
^ p(y)f{x\y) p{y)f{i\y) yeY
= max ] £ p(y)f(x\y) \ . Z€ y£Y JJ
According to the statements the parametric principle of maximal ef ficiency satisfies axioms 1-3, 16, 18 and, therefore, along with the maxi mal efficiency principle, enables one to solve Arrow's paradox of the ab sence of the principle of choice satisfying axioms 1-5 and differing from the dominance principle of the form Ndom(f) = ] J Ny, where Ny is the relation of order on R1 corresponding to f(-\y),
yeY
i.e. with i 1 , x2 € X,
The procedure of obtaining an optimal element x° € X for a specified p € [0,1] based on the maximal efficiency parametric principle is indeed the process of solving a game problem. At the same time, as in the case of the maximal efficiency principle, this problem may reduce to a maximin
Major Problems of Multiobjective
Optimization
329
problem of the following form: derive 6°, q° G [0,1], xp[S°,q0] € X from the condition 0 [pS° + + (1 - p)q p)q°] } =
max
\
*»/(«|y)>< vwweeyy
min
[pS + (1 - p)q] p)g] > •
nf(x\v)
Lemma 37.7: We have the inequality p6° + (1 - p)g° > fip. Lemma 37.8: The inequality [pS° [p£° + -I- (1 - p)q°] < < p° pp is valid. Theorem 38.4: A set of the elements Xp C X optimal by the maximal efficiency parametric principle coincides with the set of optimal elements Xp(8°,q°) for a maximin problem. The proof proceeds from the statements of Lemmas 37.7 and 37.8. The set Xp(6°,q°) of optimal elements may be presented as X°p(6°, q°) = {x°€X:6°<
fif(x°\y)
= lx° € X : min fif(x°\y)
< q°
Vy € Y}
= 6°maxfi 6° maxfif(x°\y) f(x°\y)
= = q°\.
Hence the set Xp of the optimal elements by the maximal efficiency para metric principle appears as X°p = {xeX:S°< {x E X : 6° < fif(x\y)
< q°
Wy € Y) .
The procedure of reducing the maximin problem to the extremal enables one to present [p6° + (1 - p)q°] in the following form [p60 + (l-p)q°]
= \p=
max
l\\\< l\\\
+
(l-p)qVq<EQ(x,6)}
xeX,M/(i|y)>*VyGV xeX,pf(x\y)>6Vy€Y
where Q(x,6) = {qeR11 Q(x,6) = {qeR 1 ^{qeR ^{qGR1
:6
I, max. fif(x\y) < q} 1,max.fi < q} y£Y f(x\y)
:6
Vy 6e Y} .
330
Systems Optimization Methodology
In this case a set Q(x,6) is nonempty because q - 1 belongs to Q(x, any x € X, <5 6 [0,1]. Theorem 38.5: The maximal efficiency parametric principle is equivalent to an extremal problem, thus p° = A° with a specified p £ [0,1]. The proof of the theorem follows from the fact that according to Lemmas 37.7 and 37.8 there is an inequality p° = p6° + (1 — p)q°. The maximal efficiency parametric principle makes it possible to implement parametric characterization of a set of optimal elements which might be termed a single-parameter tradeoff. Define the domain of a singleparameter tradeoff {Rp} C X of the optimal elements x° € X in the following way: {Rp} = {Xp,Q < p < 1}, where X° is a set of solutions for the original problem of obtaining fj,°p or an equivalent extremal problem. Practical application of the domain of a single-parameter tradeoff {Rp} C X is much more suitable and simpler than that of the tradeoff domain R in that {Rp} has parametrization by scalar p £ [0,1], whereas R is parametrized on the m-dimensional unit simplex in the vector optimization problem (Y contains m elements) or on a space of pseudoprobabilistic distributions on Y. The characterization method for the multiobjective optimization problem on the basis of the maximal efficiency principle which coincides with the parametric principle with p — 1 is formulated above. The characterization process based on the maximal efficiency parametric principle is carried out either for the fixed p 6 [0,1] or for all p € [0,1] in the way somewhat analogous to the foregoing. In this case it is possible to formulate and prove8 a theorem analogous to theorem (37.15) for the maximal efficiency principle. Parametric characterization supplements characterization based on the maximal efficiency principle in that an increase in the ("guaranteed" from below) degree of 6°(fi°) is naturally related to a decrease in the ("determined" from above) degree of q° of achieving an optimum with an increase of T e T for the fixed p e [0,1], here for different p € [0,1] we obtain a characterization "tube' 7 of values of <5°, q® narrowing as T € T increases.
Chapter 10 THE STUDY OF IMPROVABILITY AND PRIORITY ISSUES IN MULTIOBJECTIVE OPTIMIZATION PROBLEMS
§39. THE STUDY OF THE PROBLEM OF FEASIBLE SOLUTION IMPROVABILITY Assume that X is the set from a space X of the arbitrary elements x
identified as solutions and a vector of estimation criteria /(x) = (fi(x) /2(x),..., fm(x)) is specified on this set, moreover, each component fi( i £ I = {!,...,m}, x £ X is a unique representation of the space X in the space Rl and expressed in one scale. Assume that all estimation criteria are specified in a positive ingredient, i.e. when choosing solutions of x € X one should try to increase their values. To study the concept of comparing solutions on the set X, define the relation of order > on a space m
of m-dimensional vectors. Definition 39.1: The solution x° € X is better than the solution x1 e X according to the vector indicator / and the relation of order > on Rm if m
the conditions f(x°) > /(x1) «=» ^(x0) > f^x1) Vi e / are fulfilled. m
Define one more approach to the preference relation concept. Definition 39.2: The solution x° € X is assumed to be better than the solution x1 € X according to the vector indicator / if the inequalities
331
332
Systems
Optimization
fi(x°) > fii*1),
Methodology
1 1 fi(x°) fJ(x°)>f > Mx](x ))
(39.1)
are fulfilled for all i and if only for one j from / . We call this relation a simple order of preference and denote it by >-. It m
may be assumed, without restricting generality, that the condition should be fulfilled if x° >- xl {x° £ X is better than x1 € X in the sense of the definition (39.2). Thus x° > xl (x° is better than x1 in the sense of the m
definition 39.1), or in other words, (>-) C ( > ) • Hence of interest is the question whether for an arbitrary solution x1 6 X there is or there isn't x° in X such that x° y x1, the inequalim
ties (39.1) are fulfilled. Thus we have:
Definition 39.3: The solution x° € X is called improvable on the set X by the vector indicator / , if there exists a solution x1 £ X such that x 1 y x° m
Assume that each of the estimation criteria /; approaches extreme values on the set X. Theorem 39.1: The solution x° € X is improvable by the vector indicator / on the set X if and only if there exists an m-dimensional vector A = ( A i , . . . , Am) e Rm such that the inequalities A °)
/ J -(i°)
(39.2)
hold for all i and if only for one j from / , where the values at{\) have the form a.i(\) = a(A) — A;, a(A) = maxmin[/i(x) + A,]. We set forth the statements allowing a restriction of a set of coefficients A without violating validity of theorem (39.1). Theorem 39.2: For A € Rm to satisfy the inequalities (39.2) with a certain x° € X it should satisfy the inequalities max/fi(x) max 1 (a;) - minfj(x) for all i, j G / .
> Xj \j — X Ait
(39.3) (39.3)
The Study of Improv ability and Priority Issues in . . .
Define a set of the form A = {A £€ Rm Aj *j -- \,yi,j K,Vi,j
333
: max/j(x) max/ t (x) - min/;(x) > i g€ X
I, here here A! = 0. 0. ££ I, Aj =
x€X x£X
T h e o r e m 39.3: If for a certain A from Rm and a certain x° £ X the in equalities (39.2) are true, then there exists A' € A such that the inequalities (39.2) hold for x° and A' T h e o r e m 39.4: In the case of two components of the indicator f(x), a set A under conditions of the preceding theorem may be defined as follows: A = {(Ax, A2) G R2 : h{x2) - f2(x2) < A2 - Ai < h(x') - / 2 (x>)}, here x\ x2 £ X are obtained from the condition (i = 1,2) fi(xl) = m a x / i ( i ) . x£X
Definition 39.4: The solution a;0 £ X is called unimprovable (Paretooptimal) on the set by the vector indicator f(x) if there is no solution x1 € X which is better than the solution x° by / in the sense of the definition 39.2, i.e. there is no x1 : fix1) y f(x°). m
Theorem 39.5: The solution x° € X is unimprovable (Paretooptimal) by the vector indicator / if and only if the inequalities (39.2) of theorem (39.1) are inconsistent for any m-dimensional vector A € Rm Corollary 39.1: The statement of theorem (39.5) remains valid if the m-dimensional vector A = ( A i , . . . , Am) is chosen from the region A. T h e o r e m 39.6: The solution x° e X is unimprovable (Pareto-optimal) by the vector indicator f(x) if and only if there is an m-dimensional vector A £ A C Rm such that the equality min[/ t (a; 0 ) + A,] = max \I min[/ mia[fi(x) + A;] At] i> t (a;) + iei xex II i€i ii iel >£X t€/
(39.4)
exists. T h e o r e m 39.7: If for x° G X, A0 € A C Rm the equality (39.4) is fulfilled, then i ° is unimprovable (Pareto-optimal) and the inequalities /i(x°) > ai(A°), i = l,...,m exist.
334
Systems
Optimization
Methodology
Theorem 39.8: There are no feasible solutions of x° £ X and an mdimensional vector A0 £ A C Rm for which the following inequalities are fulfilled 0 (39.5) /.(z ) > OI i(A°), (A 0 ), 1i == 1l , . . . , m . (39.5) f J (x°)>a The statements of the theorems (39.1), (39.5), (39.7), (39.8) set a bound ary (from above or from below, respectively) for the values of compo nents of the vector indicator / by the values of the magnitudes a, (A) with A 6 A C Rm. Theorem 39.9: For the feasible solution x° G X to be improvable (Paretooptimal) by the vector indicator / , it is essential and sufficient that the following inequalities (i G I) might be fulfilled (inconsistent): fi(x°) < £ max < max min(/j(i) + A,) — A^ > A6A [ i £ X L j£l
J
J
The latter inequalities are not only of paramount importance for "con structively'1 testing the possibility of improvability or Pareto-optimality of the solutions x° E X, but also enable one to set limits on the possibility to improve the values of the multiobjective indicator components because the right-hand sides of these inequalities do not depend on x° and are de fined only for a concrete specification of an index i of the vector indicator component /, the set X and the indicator / itself here A is itself defined merely by / and X. The proofs of the foregoing statements are given in the author's work.8 Suppose that k{Xl) = {\{x) = (Ai(z),..., Xm(x)) Vi G X1 : \(x) fi{x) - fife) Vi G 7} then we have A(A^) C A for any X 1 C X.
=
Theorem 39.10: For any A G Rm there exists A1 £ A(X : ) such that ai(X1) = ai(\) for all i € J. The statement of the latter theorem enables one to substitute a set A by A(X) without violating validity of the theorem (39.3). Denote [-fi(x)] through f~(x) for all x G X and alii G I : a~ (X) = maxmin[Ai-/ l (:c)]-A,. Definition 39.5: The solution x" is termed the dual solution on the set X by the vector indicator f(x) if there exists A £ Rm such that fi(x*)>aT{X), /i(**)>oT(A), for all i and if only for one j from I.
>a~(X) f/,(**) J(x*)>a-(X)
(39.6)
The Study of Improvability
and Priority Issues in . . .
335
If we denote by A 0 a set of improvable solutions from X : A* is a set of dual solutions from X, it is evident that A * n A ° = 0 , I * U l o C A, A e = X\X°, A'* C Xe here Xe is a set of efficient (Pareto-optimal) solutions of the form Xe = {x G X\ 3xl G X : fix1) fc /(a;)}. Assume that for a given m
vector indicator f(x) and a set A C X the sets A 0 and A* are nonempty. Thus we have Theorem 39.11: The inequalities (39.2) are fulfilled for x° and A(x*), moreover, x° G A, x* G A'* if and only if for x* and A(x°) the inequalities (39.6) are completed. Theorem 39.12: For any x° G A there exists A € A(A*) such that the inequalities (39.2) are fulfilled. Theorem 39.13: The set A* is empty if and only if the set A 0 is empty. Theorem 39.14: For the solution x° G A to be improvable by the vector indicator f(x) it is essential and sufficient that there might exist the dual element x' G X such that the inequalities (39.2) would be fulfilled with A = A(x*). Consider other approaches to the issue of solution improvability. R
Define the sets 1%,..., IR SO that (J IT = I. Denote by /* r '(x) the vector indicator ordered according to an increase in indices which is obtained from the components corresponding to the indices from a set IT. Definition 39.6: The solution x' € X is better than the solution x° G A by the vector indicator f(x) if the inequalities
(r (r f/(T\x)V)>/ > /
f^(x') / ( C V) > > / (cf^(x°) V) ■*c
(39.7) (39.7)
are fulfilled for all r and if only for one c from { 1 , 2 , . . . , R}. The defini tion 39.6 is equivalent to the definition 39.2 with R = 1, i i = 1, Definition 39.7: The solution x° G A is called improvable by the vector indicator / ( x ) if there exists the solution x' G A such that the conditions of (39.7) are fulfilled.
336
Systems
Optimization
Methodology
Theorem 39.15: For the solution x° G X to be improvable by the vector indicator f(x) on the set X, it is essential that it might be improvable by / ( r ) (:r) for a certain r G {1,2, ...,R}. Definition 39.8: The solution x' G X is assumed to be better than the solution x° 6 X by the vector indicator f(x) with the priority relation {/, >) if there exists
iei:f3{x')>f3{x°)
Vie/i-
Here, on a set of estimation criteria, we specify the lexicographic relation of priority {/, >}, suppose It = {j € I : j fr i}, Definition 39.9: The solution i ° £ X is called improvable on the set X by the vector indicator f(x) with the priority relation (/,>) if there exist the solutions x' G X and i € J such that fj(x') > fj(x°) Vj G /,. Theorem 39.16: The solution x° G J*sT is improvable by f(x) with the priority relation (I, >) if there exists such /; C / that it is improvable by the vector indicator / ( l ' ( x ) . Consider the problem of solution improvability with respect to the multiobjective conceptual indicator. As stated above, we assume that there is the set X from set X of the arbitrary elements x termed solutions, and also there is the set Y from the set Y, and a multiobjective indicator is specified which is the functional defined on the set X x Y with the values f(x\y). Let f(x) = {f(x\y)}, y G Y, T = {f(x) : x G X}. Employ the order of preference (^",>) for defining the concept of a better solution on the set X. Definition 39.10: The solution x° is assumed to be better than the solu tion x' by the multiobjective indicator / and the order of preference (T,>), if f(x°) > }{x'). The following approach is also conceivable in defining the order of preference. Definition 39.11: The solution x° is assumed to be better than the solu tion x' by / if for all y and if only for one y' from Y the inequalities
/(s°b) >/(*%), f(Av) > HAV) are completed.
f(x°\y') f(x'\y') /(*V) >> /(*V)
(39.8) (39-8)
The Study of Improvability
and Priority Issues in . . .
337
This relation is termed the simple order of preference and it is denoted by {-^jfc}- It may be assumed, without limiting generality, that for any y
order of preference the relation (y) C (t>) has to be fulfilled. Of vital importance for an arbitrary solution is the issue of testing whether it can or cannot be improved by a simple relation of preference. Definition 39.12: The solution x' is assumed to be improvable on the set X by / if there exists x° from X such that inequalities (39.8) are fulfilled. Definition 39.13: The solution x' is assumed to be unimprovable (Paretooptimal) on the set X by the multiobjective indicator / if there is no x° € X such that inequalities (39.8) are fulfilled. Let f{x\y) be restricted by x and y, the sets A' and Y being compact. Theorem 39.17: The solution x' is improvable by / on X if and only if there exists the functional A = {X(y)\y € Y} defined on the set Y such that the inequalities are fulfilled for all y and if only for one y' from Y: f(x'\y)
f(x'\y')
a(y\\) = a(X) - X(y), a(y\X) X(y),
(399) (39.9)
a(X) = maxmin[ max.min[ f(x\y) + X(y)]. x€A yEY
We set forth the statements refining the set of choice of functionals A and retaining the validity of the latter theorem. Thus we have: Theorem 39.18: If for a particular A' € {Y —► R1} and a certain x' € X in equalities (39.9) are fulfilled, then the following inequalities are fulfilled for all y and y' from Y: maxf(x\y) max f(x\y) - min f{x\y') f(x\y') > X{y') X(y') - X(y). X(y). Define the following set A = {A € {Y -> R1} : max max/(x|y) f(x\y) -~ min f(x\y°) >X(y0)-X(y)Vy,y°eY}. )-X(y)Vy,y0EY}.
(39.10)
338
Systems Optimization
Methodology
Theorem 39.19: If for a particular A e A C {Y -» R1} and x' € X in equalities (39.9) are fulfilled, then there exists A' 6 A' such that inequalities (39.9) are fulfilled for x' and A' where A' = {A 6 {Y -> R1} : X(y') = 0}, max/(x|j/) - /(x'|y°) > % ° ) - X(y) Vy,y° € 7 . ISA
Theorem 39.20: The solution x' is unimprovable by the multiobjective indicator / on the set X only when there exists the functional \{y) denned on the set Y such that (39.11)
mm[f(x'\y) + X(y)] \(y)] = = maxmin[/(x|y) maxmm[f(x\y) + + X(y)}. X(y)}.
The conditions of the theorem (39.20) are sufficient if f(x\y) is concave by x for all y. Theorem 39.21: The solution x° G X is improvable (unimprovable) by the multiobjective indicator / on the set X if and only if the following inequalities prevail (are inconsistent) with y € Y: 0 f(x°\y) /(x°| ¥ ) < < maxmaxj maxmaxf min [f(x\y°) + X(y \(y0)} )} -
ASA ISA xSA I.K yy°£Y °sy
X(y)\, JJ
which are equivalent to the following equality: 0 max{ f(x°\y) - max max [ min [f(x\y°) max{/(x°|y) \f(x\y°) + + X(y \(y°)} )} - X(y)\ X(y)] } = 0 .
y£Y I S/Sr
ASA izGA 6 X Lj,°gy
JJ
Definition 39.14: The solution x' £ X is identified as active if there exists x° from X such that for all y and if only for one y' from Y the inequalities /(*'!») > f(x°\y), fix'W) > f(Av') are fulfilled. Theorem 39.22: The solution x' € X is the active solution on the set X by / if and only if there exists the functional A = {\(y)\y 6 Y} € {F —► i? 1 } defined on the set y such that for all y and if only for one y' from Y the following conditions are fulfilled f(x'\y) > a~(y), /f(x'\y')>a-(y'), ( * V ) > a - (|/'),
a~(y) = = a~ a-(y) a - ++A (\(y), j/),
maxmin[-A(j/) aa~ = =m a x m i n [ - A ( y )- - //(x|j/)]. (i|i/)]. x€X x€A y es y
(39.12)
The Study of Improvability
and Priority Issues in . . .
339
The active solutions are improvable by the multiobjective indicator with the negative ingredient of the criterion. All the above statements can be reformulated for the case of the negative ingredient of the multiobjective indicator components. Suppose we have the following subsets of the set X: improvable AY, unimprovable XH , those of the active Xa solutions on X by the multiob jective indicator / . Thus we have Theorem 39.23: Among the sets, the following relations are fulfilled Xa = 0 <^=> XY = 0 ;
XH n XY = 0 ,
XH\JXY=X,
xacxH
Introduce the representation of X in G where G is the totality of all subsets {Y —► i? 1 } placing in correspondence for each i ' € X the set A(x') of the functionals \(y) for which the inequalities (39.9) are fulfilled. Theorem 39.24:
The condition f(x)
^ f(x')
is fulfilled only when
Y
A{x) C A(x').
Denote by A(x'), B+(x'), B~(x') the representations of X in G placing in correspondence for each x' from X the sets of the functionals \(y) for which inequalities (39.12), equalities (39.11) and the following equalities are fulfilled max[f(x'\y) yEY
+ \(y)] = minmax[/(a;|y) + X(y)]. x€X xEX
yEY y£Y
Let B{x) = B+(x) n B-{x), denote for X' C X by A(X') the image of the set X' with the representation A(i), x € X' similar to A(X'), B(X'). Definition 39.15: The set A C G is termed sufficient provided A(X) n A = 0 , if and only if A(X) = 0. Theorem 39.25: The set A = B(Xa)
is sufficient.
Definition 39.16: The set A C G is called a-sufficient provided A(x)C\A = 0 , if and only if A(x) = 0 with x € X. Theorem 39.26: The set A = B(XY)
is a-sufficient.
340
Systems
Optimization
Methodology
Now consider the case of arbitrary X and Y, having rejected the com pactness assumption. Definition 39.17: The solution x' 6 X is identified as e-improvable on the set X by / if there exist x° e X and e > 0 such that the inequalities are fulfilled for all y and if only for one y' from Y: f(x'\y) f(x'\y) < f(x°\y) + e ,
f(x'\y')
< f(x°\y') + e .
Definition 39.18: The solution x' £ X is referred to as improvable on the set X by / if there exists e > 0 such that x' is e-improvable. Theorem 39.27: The solution x' € X is improvable by / on the set X (in the sense of the definition 39.18) if and only if there exists the functional of A = {X(y)}, y &Y such that for all y and if only for one y' from Y f(x'\y)
f(x'\y')
The latter relations may be written in the following, sometimes more suitable, form for all y, y' and if only for single y, y' from Y: maxf(x\y)-f(x'\y')>X(y')-X(y), x£X
maxf(x\y)-f(x'\y')>X(y')-X(y). XfzX X GA
Now suppose on the set T = {f(x) : x G X} several orders of preference (•T , ^ i ) , (-^1^2), •••, {F, 5u) are specified. 7
Definition 39.19: The solution x' € X is called improvable on the set X by the multiobjective indicator / and a set of the preference orders {!>i. >2, • • •, ^z} specified on T if there exists x from X such that for all / and if only for one /' from { 1 , 2 , . . . , Z} there exists f(x)>if(x'), f(x)>if{x'),
f(x)»t,f(x').
The Study of Improvability
and Priority Issues in . . .
341
Now suppose on the set Y, the lexicographical relation of order (Y, >), called the priority relation, is specified. Denote Yy = {y' e Y : y' > y}, where y € Y Definition 39.20: The solution x' £ X is better than the solution x° € X by the multiobjective indicator / with the specified priority relation (Y, >) if there exists y° from Y such that x' is better than x° by T{x\y°) and a simple order of preference on Yyo where
y'eY y'eY0
T(x\y°) = {f(x\y')}, T(x\y°) {f(x\y')} ,
It may be shown that the introduced relation of preference is asymme tric, transitive and antirefiexive. Definition 39.21: The solution x' € X is better than the solution x° e X by the multiobjective indicator / if there exists a complimentary unique representation p = {p{y)}, y € Y of the set Y in Y such that for all y and if only for one y' from Y there exists f(x'\y)
> f(x\ p(y)),
f(x'\y')
> f(x\
p(y')).
The above relation of preference is asymmetric, transitive and antire fiexive. Definition 39.22: The solution x° e X is improvable by the preference relation introduced in the last-mentioned definition on the components of the multiobjective indicator / on the set X if there exist y' e F and the solution x' € X which is better than x° G X according to this preference relation. Denote for the fixed x' £ X; y, y' 6 Y; Xyyi = {x £ X : f(x\y) > f(x'\y')} and define the content of these sets as \Xy>\ = U n Xyyi, where U is the set of all subsets of Y. Theorem 39.28: For the solution x' G X to be improvable in the sense of definition 39.22, it is essential that there might exist y' E Y such that Suppose we have the following representation of Y in Y: q(x,x'\y)
e£ {y' E € Y*(y) : f(x\y')
=
min
y"€Y'(y)
f(x\y")} /(*[*")},
342
Systems
Optimization
Methodology
here the set Y*{y) is defined as: Y*(y) = W £Y:
f(x'\y) > f(x\y')}\q(x,x'\
Y'(y) = {y' £Y:
f(x\y') <
= |J
Y'(y)),
f(x\y)},q(x,x'\Y')
q(x,x'\y).
y£Y'
Theorem 39.29: The solution x' 6 X is better than the solution x £ X if and only if for all y and for at least one y' from Y f(x'\y)>f(x\q(x,x'\y)),
f(x'\y')>f(x\q(x,x'\y')).
The foregoing definitions and statements form merely part of the di versity that occurs owing to employment of the simple preference relation on the components of the multiobjective indicator / on the set T for ob taining the preference relations of one solution as compared with the other as well as for formulating different approaches to defining the concepts of improvability and generalized Pareto-optimality (unimprovability) and the necessary and sufficient conditions of improvability or optimality according to the preference relations introduced. Consider the problem of improvability depending on the optimum achie vement degree. Instead of the multiobjective indicator / in the definitions of improv ability of the solution i ° 6 A' we shall utilize Hj(x) = {fif{x\y)}, y € Y the components of which are the magnitudes n;{x\y) is the degree of achieving an optimum at the point x £ X by the yth component f{x\y), thus we get: A X(y°) - X(y)\ < l.Vj/.s, l,VV,y° 0 £€ Y) Y) K , = {A e {Y - R1} : | \(y°)
(39.13)
Definition 39.23: The solution x° € X is improvable according to the multiobjective indicator fif on the set X if there exists a solution x' 6 X which is better than the solution x° £ X in that for all y £ Y and if only for one y' € Y the following inequalities are completed M/(*'W > Hi(x»\y), »f(x'\y)>Hf(x°\y),
nf(x'\y') f(x°\y'). > ^f(x°\y'). Hf(x'\y')>ti
(39.14)
Definition 39.24: The solution x° 6 X cannot be improved (Paretooptimal) according to the multiobjective indicator ptj on the set X if there
The Study
of Improvability
and Priority
Issues
in . . .
343
exists a solution x' £ X which is better than x° £ X in the sense of these inequalities. Theorem 39.30: The solution x° £ X is improvable (unimprovable) ac cording to the multiobjective indicator ^f if and only if there prevails (are inconsistent) the following inequalities with y £ Y: M/(x°|j/) Hf(x°\y) < maxmax{min[ M / (a;|2/ 0 ) + A(2/ 0 )]-A(2/)}, which are equivalent to the following equality: max\nf(x°\y)ms^\iif(x°\y)yEY
k
0 max max min [n [fif(x\y (x\y°) )
+ \(y0)} - \(y)] A(f)l } = 0 .
A€A M xeX i € A Ly°€.Y Lj/°eY
J J
In the same way we may also reduce the conditions for improvability of the solutions x° £ X or their Pareto-optimality by using \i^ instead of / . Definition 39.25: The element x* £ X is referred to as superoptimal according to the generalized principle of maximal efficiency if the conditions are fulfilled for all y £ Y: tif(x*\y)
= AeA max{i j min [ f * / ( * V ) + AHy°)} (^°)] i°gy M
Kv)\ Kv)} )
0 0 0 = maxmax{min[/x )]-A(y)) maxmax{min[/x//(x|i; (x|T;)-l-A(iy )-l-A(/)]-A(j;)).
The superoptimal solution set AT* is the contraction limiting by A € AM not only of the Pareto-optimal solution set XH but also of the maximal efficiency criterion-optimal element set X°, i.e. AT* C X° C X°H C X. On the one hand, it enables one to carry out characterization of the prob lem of the / multiobjective optimum on the set X, using the generalized principle of maximal efficiency as the base, and formulate the parametric generalized principle of maximal efficiency, on the other. In this case the values fif(x*\y) with y £ Y constitute the limiting bounds for possibilities to improve the solution of x° £ X (in that x° £ X can be improved if and only if fif(x°\y) < fif(x*\y) for all y £ Y) as well as the limiting bounds for the best choice of x° £ X% (Pareto-optimal) by specifying conditions for x° £ X*.
344
Systems Optimization Methodology
§40. THE STUDY OF THE PRIORITY PROBLEM IN MULTIOBJECTIVE OPTIMIZATION Consider the problem of priority in multiobjective optimization. Priority in the problem of optimal choice of an element x° € X by the multiobjective indicator /(x) = { f ( x \ y ) } , y G Y, where Y is a set of objective terms with the fixed X C X, Y C Y, f € F is taken as specification of the priority relation on the set Y according to importance, on the components /(• \y) of the multiobjective indicator / with y € Y according to significance (preference), and also on the values of f ( x ) £ f of the multiobjective indicator / with x £ X according to efficiency (a degree of increase in the values of /). The priority relation may also be specified on the elements x £ X themselves, but in this case it is understood as specification of the principle of choice. In multiobjective optimization problems, the priority on X (i.e. the principle of choice) generally is not specified, because, if it is not specified, then there is no problem of deriving the principle of choice, and there remains the calculative problem of deriving
optimal elements. If there is no specification of priority on Y, f(x) with x £ .Y, T and priority is specified on X as the principle of choice, then it remains to obtain a solution of an appropriate extremal or maximin problem. Hence the problem of taking into account a specified priority aggravates solving the problem of deriving the principle of choice which satisfies a specified or unspecified totality of axioms and properties for which this priority should be fulfilled on optimal elements. The theory of multiobjective optimization is still in the making, hence it is clear that the problem of priority is complex and differently understood by various authors. A common approach to the problem of definition, specification and incorporation of priority is still lacking. Keeping this in mind, one should not anticipate an exhaustive solution of the problem. We focus on the following three aspects of the concept of priority in multiobjective optimization: priority on the multiobjective terms of y G Y, on the multiobjective indicator components /(• | y) with y £ Y and on the values f ( x ) € T of the multiobjective indicator /. Definition of the priority relation is based on the comparison of the objective terms y e Y in the scale Cy of the importance of y € Y, the components of /(x|y) with x € X for the fixed y € Y of the multiobjective indicator / in the scale Cf(. | y) of significances
of the quantitative values of f(x\y) of the y quality being obtained when specifying x € X, and the values of f ( x ) 6 T in the efficiency scale Cf for
The Study of Improvability and Priority Issues in ... 345 a totality of the quantitative values of {/(a%)}, y € Y of all qualities of y € Y obtained when choosing x € X. The scale Cy is identified as a particular order or a list of indices on a specified set Y of the objective terms (expressing different qualities, characteristics, properties, etc.) generated by specifying the elements x e X. This order determines the importance either among objective terms or objective terms and subsets from Y or subsets from Y. When comparing objective terms or subsets from Y the importance is taken as one or several estimates of utility (values, etc.) of the qualities under consideration corresponding to objective terms y from Y. Sometimes this order indicating the importance of objective terms in the scale Cy is not given and, in compliance with it, with the index enumeration of objective terms y from y the partitions of a set of indices by the importance (which may not be specified, though related to its quality and every quality is related to its index. In a number of instances that kind of order is specified as mixed, i.e. when comparing some subsets by the importance and giving indices or their subsets. But here we need to correlate specification of the importance of subsets of objective terms as qualities and specification of the importance of these subsets for terms as indices. The scale Cf(. \y) is taken as a total of the obtained various degrees (gradations) of the y quality in the quantity f(x\y) of different elements x € X, the ordering on which is termed the significance (preference) of obtaining some quantitative values of the quality of /(• | y) as against the others. If y is a totality of quality indices as objective terms and there is no indication in the scale Cy as to the importance when comparing different objective terms y from Y, then the significance in the scale Cj{. \y) can be presented as the importance of obtaining some quantitative values of the y quality as against the others. Moreover, here the importance of obtaining quantitative values of /(• \y) from a particular set can be specified when comparing them with a set of quantitative values of /(• \y) of another quality y'?y. LetC/ = {C / ( .| s ) },i/€y. The scale Cf is taken as specification of a natural order Ndom on the multiobjective indicator values of f(x) in the following form for x1 ,x2 € X: /(x^N^fix2) <=» f(xl) >/(z2) ^ f(xl\y) > f(x*\y) Vy € Y,
346
Systems
Optimization
Methodology
or the relation of order jVdom of the form f(x1)Ndomf(x2)^>{[f(x1)Nd°mf(x2)} l 2 0 A[3y°£Y:f(x \y°)>f(x A[3y0€K : / ( * V\y) )}} >/(*V)]}
Specification of the set X of admissible elements is supplemented, in a number of problems with constraints on the multiobjective indicator values, though, basically, it leads to substitution of A' by its particular subset A" C X. When comparing the objective terms of y 6 Y, the components f{\y) of the multiobjective indicator /(■) = {/(■ | y)}, the values of f(x) with x £ X of the multiobjective indicator / £ F from the subset T — {f(x)}, x £ X there may prevail equivalence of the objective terms of y, the components /(■ | y) and the values of f(x) in their scales Cy, C/(. | H ), Cjr respectively. In this case, the equivalence y1 «■> y2 of the objective terms y1, y2 £ Y C-r
in compliance with the scale Cy is understood as the situation when the importance of the objective terms is the same or each of them enters into subsets of >j and Y2 C Y The equivalence of Y\ and Y2 C Y in this case signifies the same importance of a generalized quality of l'i, Y2 as a totality of qualities expressed by the objective terms y £ Y\, f £ Y%. By the equivalence of f(x\yl) ** f{x\y2) with y1 ^ y2 £ Y for a particular element x € A" we mean fulfillment of the following relations: 1 fixly ) 1 )
f(x\y2) ^ {[y1 r y2)V (y1 py2)} 2 y 2 f(x\y ) ^ {[y1 rL-Y y2)V (y1 C py )} r A 2 w CC y l C Cf (/(■ U1) r L-Y C f(-\y )) K f(-\y )P f(-\y2)) A w C (/(■ U1) Lf r C / ( - |y2)) K f(-\yLlf)PCf(-\y2)) 2 1 1 AK/Cxlt/ A"22),), ( # , AA' AK2)2) C (7/]]} (?/]]} , AKfixly ) ) e ^A^)A(/(;r|2/ ) A ( / ( a : | s 2 ) 6 A 2 ) : ((A' # ix - A r
Lf
i.e. /(xl?/ 1 ) is equivalent to /(a;|y 2 ) with y1 7^ y 2 , y 1 , y 2 £ Y if and only if y1 and 1/2 are equivalent and incomparable in the scale Cy, the scales Cf(. 1 j,i) and Cj(. | y2-j are equivalent and incomparable in the scale C/ and fi^ly1), f{x\y2) belong respectively to the sets kx,k2 € C/ equivalent by the scale C/, where by D we denote the binary relation of incomparatibility. The equivalence / ( i 1 ) ^ /(x 2 ) corresponds to fulfillment of the following
The Study
of Improvability
and Priority
Issues
in . . .
347
relations: fix1)
r
fix2) f(x
<=► {[/(x^N^fix {[fix^N^fix2)
A /(*»)JV**7(* 1 )]
V[/(x1)iVdom/(x2)A/(x2)7Vdom/(x1)]}, in other words, the values fix1), fix2) € J" are equivalent in the scale Cf if and only if these values are equivalent either by the relation of order iV dom or by Ndom An example of equivalence is provided by y1 = y2, 1 1 2 2 / ( x IJ/ ) = f{x \y ), although x1 ^ x2 is possible. It is possible to secure another form of comparison for the objective terms y € Y, the components /(• | y) of the multiobjective indicator / and the values of /(x) £ T with x € X where unequivalence prevails, by which we mean the situation when at least one of the forms of equivalence is absent in the scales Cy, Cf, Cjr. In this case (of unequivalence), when specifying y1 ^ y2 6 Y, x1 / x 2 € X and / £ F we define relations > , CY
> and > in the following way: y1 > y2 if y1 is more important than y2 Cf
CV
CY
in the scale Cy; /(• |y 1 ) > / ( • | y2) if the ?/th component of /(• \yl) of the multiobjective indicator / has more significance (preference) in the scale Cf as against the y 2 th component of f(-\y2); fix1) > f{x2) if the value of fix1) of the multiobjective indicator / on the element x 1 e X exhibits more efficiency as against the value of / ( x 2 ) on the element x 2 6 X. The binary relations of preference > on the scales Cy, Cf, C? enable one to introduce similar relations between an element and a combination of elements from Y, {/(■ | y)},y € Y, T combinations of elements of these sets both separately by every scale and jointly. Thus we have Yj>> 2/2,2/i Y j/2,2/i > « , { / ( * ) } , € * ! > CY
CY
c?
fix22)---
We consider now the definition of priority and briefly discuss description of methods for its specification. At the outset we have the case of specifying the two-component multiobjective indicator and the two-element set X, i.e. Fi = {yx,y2}, Xi = { x y x 2 } for some y1 ^ y2 € Y and x1 ^ x2 S X. Suppose the scales Cy, Cf, C? are given with some specified relations of importance > , significance > and efficiency > , and if it is clear which CY
Cf
Cy
scale is utilized, then it will be omitted under the sign >. For the elements y1 ^z y2 g Y in the scale Cy by the binary relation Cy one and only one
348
Systems
Optimization
Methodology
of the following relations can exist: y1 >y2, y2t>y1, yl -* y2, y1 >y2, y2 >yl, y1 ^y2, yxOy2 For components /(• | yl), /(• | y2) of the multiobjective indicator / € F in the scale Cf by the binary relation > it is possible to fulfill only one of the following relations: 2 /f(-\y ( ■ 1| )>f(-\y y 1 ) > /2()J(-\y*)>f('\y - | y 2 ) , / ( - | y1)J(-\y ) > / 1()-f(-\y ' | y 1 )2,)i/ ( - | y 1 ) - / ( - | y 2 ) , 1 1 2 22 1 1 2 1 1 2 /f(-\y ( • | )>f(-\y y ) > /)J(-\y ( - | y)>f(-\y ) , / ( )J{-\y ' | y 2 ))^f(-\y » / ( - |)J(-\y y ) , )Of{-\y / ( - | y 1).) - / ( - | y 2 ) , / ( - | y 1 ) n / ( - | y 2 ) -
In the case with two values of Tx = {/(x 1 ly 1 ), fix1 \y2)}, T2 = {f(x2\yx), f(x \y2)} of the two-component objective indicator {/(• | y1), /(• | y2)} in the scale Cjr by the relation > only one of the following relations can be cv satisfied: T\ t> T2\ T2 > Tx; T\ ^ T%\ T\ > T2\ T2 > T\; T\ -~ T2\ T\ DT%. In this case the priority for the objective terms yx,y2 € Y is taken as a requirement to fulfill only one of the relations y1 t> y2; y2 > yl we 2
have Pri{yl,y2)
=> y1 > y2, PTt(y2,yl) CY
CY
Cy
<=> y2 > y1. Priority for the CY
components /(• | y1), f(-\y2) of the multiobjective indicator from these two components is understood as a requirement to fulfill one and only one of the relations: / ( l y 1 ) » f(-\y2),
f(-\y2)>
C, G«
Cf( C
/(-ly1).
i.e.
^.(/(■|y1),/(-|i/2))<^/(-|y1) > / ( V ) , Cf 2
11
22
^ . (( //((••| |2y/ ) , //((--l|2y/ ) ) ^ ^^//((-||yy ) > / M (ly1)Priority for the values of T\(xx) and T2(x2) of the multiobjective indi cator / is assessed as a requirement to fulfill one of the relations Ti > T2, Cjr
T2 > Tx, * & Cr* '
The Study of Improvability
and Priority Issues in . . .
349
i.e. F rr.1((JJrrii((xx11)),,^^22 ( a ; 2 ) ) < = > ^ li
> > ^2, Cr Cr
and the meaning of Pri(Jri(x1),Jr2(x2)) resides in the following. For the element x1 the value F\(xl) of the two-component multiobjective indicator displays more efficiency in the scale Cj? as against T2{x'2), for the element x2, respectively, the opposite is true. When on the scales Cy, Cf, C? the relations of importance > , sigCy
nificance > and efficiency > are specified as perfect orders, the priority Cf
Of
on the sets of objective terms Y, components of the multiobjective indi cator / and values of T is understood as specifying on them the subsets of chains (series) completely ordered by appropriate relations >, and the concept of priority specification P r , (/Ci,/C2) <$=$■ )C\ > fC2 for two subsets or P r , (/Ci, fC2,. • •, fCm) <=> {fCi > /C2 > . . . > K.m } will be defined by us ing generalization of the binary relation of priority in the requisite scale C as the base. Here specifying the relations of importance, significance and efficiency enables one to establish priority not only separately in each scale but also in a common scale Co = Cy U C / U Cy on a direct product O = Y x / x f, where / is taken as a set of the multiobjective indicator components, in other words, / = {/(• \y)},y € Y Therefore, when specifying the scale Co and the relation > the priority PTi{0, CQ} is taken as a chain (one or several) of subsets from 0 ordered by the relation > In the general case we can also place in PTi{0,Co} Co
specification of a set of relations { > }i^c, or the scales {C 0 }/ 6 £ themselves, or their part, or the part that has been specified before formulating priority in the statement of the problem. The multiobjective optimization problem is assumed to be specified if we determine the sets X C X oi feasible elements x, Y C Y of objective terms, where / € F is the objective indicator, f is the value f(x) of the multiobjective indicator with x £ X and priority is PTi{0,Co}After the multiobjective optimization problem has been formulated the problem arises with respect to determining the principle of choice and the
350
Systems Optimization
Methodology
elements optimal by this principle. This is according to the specified priority or the incorporation of the specified priority with the utilization of a principle of choice depending on which of the following assumed set of possible cases is realized: Is there the scale Cy on Yl If so, then is the order > specified by importance? CY
Is the priority on objective terms and subsets from Y specified with the use of that order and scale? Is there the scale C/ on the multiobjective indicator components / € F? If so, then is the order c> specified by importance? i Did we specify with the use of that order and scale the priority on the components of the multiobjective indicator / and the multiobjective indicators from F determined by various combinations of these components (special indicators for /)? Is there a dominant element x° € X by the relation of order Ndom according to the specified X,Y, f £ F and the priority P ri {O,Co}? If so, then x° is the multiobjective optimum. If it is not, then is there the Pareto-optimal element x° € X by the T dom
relation of order TV according to the specified X,Y,f r
The Study of Improvability
and Priority Issues in . . .
351
basis given by the specified relations of order. As noted above, realization of that principle is based on generalization of the principles of choice as a linear convolution with pseudoprobabilistic distributions as well as gen eralization of the maximal efficiency principle. We shall set forth a few illustrative examples of utilization of the possibilities of these approaches as applied to vector optimization problems for incorporating priority and obtaining a multiobjective optimum for several special cases. We start with incorporating priority in the maximal efficiency principle. We have the set X of feasible elements, the set Y = I = { 1 , . . . , m} of the objective term indices and the multiobjective vector indicator f(x) = {fi{x)i • ■ • i fm(x)} as a totality of quantitative values of the objective terms (qualities) corresponding to the feasible elements x £ X. Assume that the components /, of the vector indicator / have a positive ingredient (when selecting elements x £ X one should seek an increase in their values). Sup pose we do not specify the scales of importance and significance (preference) on the set Y and components of the vector indicator / , and there is no pri ority specification. There exists the possibility of introducing the efficiency scale on the basis of utilizing natural normalization of the vector indicator components as (i 6 / ) , ^i(x) = \fi{x) — ff]/[f* - /, 0 ]- Here Hi(x) is the de gree of achieving an optimum by the ith component of the vector indicator / for the element x £ X, f° = minfi(x), f* = max ft(x) and the scales of efficiency by each component \ii in a quantitative form are dimensionless and represent intervals [0,1] (for any i £ I the inequalities 0 < /xt(x) < 1 are fulfilled). For the maximal efficiency principle in the conceptual form of specification, verifications of the fulfillment of axioms as well as properties and characterization of the sets of elements optimal by this principle are given above. This enables one to transfer properly the axioms, properties and characterization onto the multiobjective vector indicator case. More over, the above properties of improvability, Pareto-optimality and many others for the feasible elements x £ X can also be properly extended to the case of the vector indicator / . Following the maximal efficiency principle, the optimal element x° £ X is obtained from solving the following maximin problem: maxmin/x^a;) = min^i(o: 0 ), the latter problem may be reduced to the extremal problem: i€l
H°(x°) = min/i 2 (x°) = t'6/
max 0<JX<1 x£X,[ii(x)>fl
i€/
fi = max [ft\in(x) > y,,Vi 6 / ] . 0
352
Systems
Optimization
Methodology
Let n(x) = (fii(x),...,fj.m(x)), the maximal efficiency principle is termed the maximal efficiency ^-criterion or simply the /x-criterion, and the corresponding maximal or extremal problem is identified as the ^-problem. Consider several specific cases of priority specification in the vector opti mization problem. (1)
Assume that on the vector indicator components f(x) = (fi(x),..., fm(x)) we specify a simple relation of priority (lexicographic order) of the form l>2>--->i>i + l>->m which does not change when utilizing the natural normalization fii(x),i € / . For that case the \icriterion is formulated as follows: u? = max u\ex s£ 0 , where via 0<M<1 x„ X1** denotes the set of elements x £ X as X%* = {x £ X : n(x) > (m — i + l)/x,i = l , m } . In so doing, the magnitude rfex has the form /j°ex = max{min[/x;(a;)/(m — i + 1)]}, here the ^-criterion optimal element x° £ X satisfies the inequalities fit{x°) > (m — i + l)M?ex-
(2)
Assume that on the vector criterion components the priority binary relation > is specified such that i > j , j £ U, (i = l , m ) , here we designate /, = {j £ {l,...,m} : i > j}. In this case the /^-criterion is formulated as /i°(>) = max u, here Xu{>) is the set of elements x G X of the form Xu.{>) = {x £ X : m{x) > (m, + + l)fi
(i = = T~m)} l,m)},
TO, is a number of elements of the set I, C I = { 1 , . . . , m } . Similarly it may be shown that the magnitude /i°(t>) takes the form /x°(t>) = max] min[^ /x°(>) min[/jI (x)/(m1 + 1)] >, where the /i-criterion optimal element x° £ X satisfies the inequalities Ht(x°) > (m, + l)/i°(>), i £ I. (3)
If priority is specified as i>Fs (s = 1,.,.,m,-,i e I), where Ps C I, the //-criterion is stated in the following way: u (>) =
max 0<M<1
a,
The Study of Improvability
and Priority Issues in . . .
353
here Xh(>) = | x E X : ^(as) ^(95) >
J2m(Fs)
+ 1 p,
iE € l\ ,
m(Ps) is a number of elements of the set I\ C 7, and the magnitude /x°(>) is equal to fj°{>) «= max Ij min 1J * ( * ) / M°(>)
£
m(F m(/*) + 1 11 s) +
here the /z-criterion optimal element x° E X satisfies the inequality mi
w£*°)>
^m(/:) + l
Ai°(>).
. s= l
Consider the possibility of applying the /j-criterion to derive the weight coefficients of linear convolution. In the vector optimization theory one of the most essential principles of optimality is the linear convolution criterion of the form m
5 M (x°|a) = ^mQ , - / / i ( x ) , ffM(x°|a)
= ^Q,/Lit(a;),
where the weight coefficients o j , . . . ,am satisfy the conditions 0 < a, < 1, {% — 1,10k), Yl a * — 1- I n those cases when the weight coefficients i= l
Q i , . . . , a m are specified, the optimal element x° E X is derived from the condition gfJ,(x°\a) = iaaxgfl(x\a). The problem of formulating the methods for obtaining weight coeffi cients still remains unsolved. Consider the application of the ^-criterion when solving this problem. In the above priority specification cases, the /x-criterion for obtaining weight coefficients is stated in the following way. (1) Specifying a simple relation of priority (the case of lexicographic order) of the form 1 > 2 t> ■ ■ ■ t> i > i + 1 > ■ ■ ■ > m, obtain 00
A= { ( a i , . . . , a m ) ER € . Rm,m , 0 < a, < < 1 (i =T~m), = T~m), £ a* a{ = = l } , and the ••
/x°ex satisfies the condition
i=l
'
354
Systems
fi°ex M° e x ==
/
Optimization
max
Methodology
{ m i n [ Q ],//( m - i + ll)]} )]|
Moreover, tfex = 2/[m(m + 1)]. In this case the /x-criterion optimal weight coefficients of,... , a ^ satisfy the conditions (i £ I) 771
0 < att?° << l1, ,
a° > 22 (( m - i + l)/[m(m + l ) ] ,
a X !X ?
=
1
'
3=1
These conditions, particularly, are satisfied by the weight coefficients of the form a° = 2(m-i + l)/[m(m + l)], (i = 1,. .. ,m), which represent the point estimates for a simple relation of priority and implement the maximin problem solution. (2) Specifying the binary relation of the priority i > j for j S U = {I £ {l,m} : / > i}, i g / , obtain u°(>) =
max
u,
0<M<1
where A^(>) = { ( a l 7 . . . ,am) € A : a; > (m t + l)/i, t S 7} and the magnitude Af0(>) is represented as M°(>) =
max
^ min[ai/(m 1 + 1)] >.
It may be shown that
/
L j=a j=i
and the //-criterion optimal weight coefficients QJ, . . . , a ^ satisfy the conditions (i £ I ) , 0 < of < 1, m
/
.3=1
m
\
J=l
The Study of Improvability
and Priority Issues in . . .
355
These conditions are satisfied, in particular, by the weight coefficients {i £ I), a° = (771; + 1) /
^ (m; + 1) which complete the maximin
problem solution. (3) Specifying priority of the form i>I\ where Ps C I, obtain u, (>) =
max
(s = l,...,ml;i
= l,...,m),
a,
0<M<1 0<M<1
^ 7 n (a/)s I ) + ^m(P + lI / an,,
AM(*)=|(ai,...,am)€i4:ai> M°(>)=
max
iI m i n i m a,/
Y
Vi € 7/ I ,
m(Ps) + I 1 1 ;
it may be shown that
I
/ r m r »,
1
]i
A>)= /{E[X>W+ j, the //-criterion optimal weight coefficients satisfy the conditions m
O
a =1 =1 a
I" m;
/r m
m
E °° >- >> E ^ ) + ! / E jj=i =i
Ls=i L=i
( mj
\
E
*
JJ// L>=i L>=i U U=i =i
■ JJ..
These conditions, in particular, are satisfied by the weight coefficients of the form *mi
"\ l (
. 8=1
J/
rn
\
n%i
11
m a°= 5>(/]) + li / E (4) + 1 r[ a°= X>co E EX>(ii)+i I j ==ll LL ««==11
JJJJ
which realize the solution of the appropriate maximin problem. The above specific cases of specifying and incorporating priority enable one to formulate general approaches to employing the maximal efficiency principle and obtaining the weight coefficients of linear convolution to solve the problem of incorporating more complex priority specification forms in multiobjective optimization problems.
Chapter 11 PROBLEMS OF MULTIOBJECTIVE OPTIMIZATION UNDER INFORMATION DEFICIENCY
§41. PROBLEMS OF DECISION MULTIOBJECTIVE OPTIMIZATION UNDER UNCERTAINTY The situation of decision making in multiobjective optimization is re-
ferred to as the pair {X, /}, here X = (x\,..., xmx) IS a se^ of decisions of the governing unit M;f(x) = {fi(x)}^=1 is a vector of the estimate functional defined on X and assuming their values from R1 In the thus defined multiobjective situation of decision making {X,/}, the decision making problem resides in the fact that the governing unit M must select one decision optimal by the principle of choice adopted by this unit. The multiobjective decision making problem is associated with three factors {u,v,w}, here u is a normalization procedure; v is a convolution criterion; w is a priority relation. In the given case the normalization procedure u is
identified as a unique representation of Rf in RL Normalization is used for substitution of an ingredient and transition to the comparable scales in the estimate functionals. The priority relation w is referred to as the vector of estimates (wi,..., WL), in particular cases priority may be specified by the relation of order. The convolution criterion v is identified as the rule determining the decision for a given situation of multiobjective decision making which may be adopted as optimal. The convolution criterion represents, as a rule, the functional mapping RL in Rl The main directions of research into the multiobjective decision making problems should involve development of the choice of the main elements of 356
Problems of Multiobjective
Optimization
Under . . .
357
convolution, the methods for normalization and priority incorporation, and development of approaches to the problems of analyzing and substantiating the choice of these elements. Suppose {X, / } is a decision making situation, where X is the set of decisions of the management unit M; f(x) is the vector estimate functional on the solution x € X. Assume that we have two sets: U = {uq,q € Q), the set of convolution criteria and V = {Vu,,w € ft}, the set of normalization procedures. An unambiguous definition of the normalization procedure and convolu tion criterion is complex and even impractical for many vector optimization problems. In our case we have several versions corresponding to various conditions for selecting an element from the set: the choice of a unique element from the set (A) is determined; none of the elements of the set is satisfactory (B); all the elements of the set are acceptable (C). In this case for the set U we have the cases Aj, Bj, C[, for V — A%, B2, C2 and for the multiobjective decision problem we have nine versions: AM2; AiB2\ AiC2\ BjA2; B,B2; B,C2; dA2; C,B2; CrC2. Let Fq{x) denote the vector estimate functional for x € X obtained after employing q normalization (q £ Q), similarly (with u> G 0) let vl(x) = vu(^Fq(u})) for x e X, vl{x"u) = opt t)«(arfc). xkex For B[ we assume that the convolution criterion is determined on the non-normalized components of the estimate functional and, therefore, the situation coincides with Aj. In the cases ^ / ^ the unchanging index q or us is omitted. Consider several approaches to formulating the problems of choosing u, v based on the guaranteed result criterion and employment of the uncertainty functions. Approaches to formulation of these problems may be divided into two groups: the first includes convolution procedures with regard to all convolution criteria (or all normalization procedures, or both according to a version); the second comprises the methods of deter mining one convolution criterion (a normalization procedure). Denote:
6i(x ) = K(xn-vi(x M(xn-vl(x )}K(xn, k)}K(xn, 6qJxk)k =
k
ri(xk) = max el(x el(xkk)) =
l — i.,...,Lt
\\x*"-xkk\\, \\x*"-x
i {[J*(*-)-f i(*-)-fi,{*k)}/fi,(**')}, r{[7
358
Systems
Optimization
Methodology
here |j ■ || denotes a number of nonzero components. Consider the formula tion procedures for the problem of choosing u, v for each version. Thus we have: In the version AjA2 a unique criterion of v convolution is used for a specified unique normalization of u when deriving xko £ X as an optimal solution from the condition v{u[f(xko)]} = max v{u[f(xk)}}. Xk€X
In the version BjA2 a unique criterion for v convolution is used when obtaining xko as an optimal solution from the condition (the estimate func tional components are not normalized) v[f(xk°)] = max
v[f(xk)].
IHJ -t G A 6A
In the version C;A2 we have a set of normalized vector estimate functionals Tq{xk) and, respectively, 6q(xk), lq{xk), eq{xk) with q £ Q,xk £ X, and the unique criterion for v convolution is utilized when obtaining a so lution to one of the following problems: (1) H(xko) = max H(xk),
H(xk) = - £
I t 6 A
rfcGX
qeQ qeQ
vq(xk)=v\T )=v\fqq(x(xkk)\/{ )}/{ L
!0
(2) A' {xko)=
£ J /
vq(xk)\nvq(xk),
v[T v[Fqq(xkk)}\)}\;
K
q€Q
'
min maxA'(xjt), here as Aq(xk)
ii-e-f qeQ
we can use 6q(xk),
fq(xk),
eq(xk) with the fixed to £ fl corresponding to the specified convolution of v; (3) H{q°) vq(xq^)\nvq{xqiu'y, H(q°) = = maxH(q), H(q) = - £ (4) A' 0 (i'"'") = minmaxA«(2:" u '). qeQqi€Q
In the versions AjB2; BiB2\ C\B2 we have respectively a unique nor malization of u[f(x)}, the non-normalized f(x) and a set of normalized Tq{x) with q £ Q, x £ X, and convolution criteria are unacceptable. In this case, in the above problems (l)-(4) the choice of normalization and an optimal solution are formulated in the same way, if formally as a criterion for v convolution we employ transformation of summing by the estimate L
functional components v[f9(xk)] — J2 ^'q{xk)In the versions AjC2\ BjC2; C[C2 we have, respectively, the unique normalization of u[f(x)], the non-normalized f(x), a set of normalized
Problems of Multiobjective
Optimization
Under . . .
359 359
Fq{x) with q £ Q, x £ X and the set v%(x) = v^\Tq(x)\ of convolution criteria with u £ ft. In the version CiC2 the problems of the optimal choice of the optimal solution according to the multiobjective estimate functional are formulated as follows: q (1) H(xko) = max maxff(a?fc), H{xk), H(x H{xkk) = - £ J2 §J(3:v^(x !*(*&), fc}lQ k)lnv u(xk), Xk€X Xk£X
q&Q
vl{x = «2(*k)/ vl{xk) «£(*k) k) /
£ «£(**) £ «£(**) ;; L
g ^00(x milimaxA j^(x), (2) A A*° (:rjt°) A«' f( ni t)),, here as A»(*) A?(x) we may use &%{x), "fc(x), ko) min Xk€X
qeQ
ffteV0)0) = = maxH{q,w),H{q,w) maxff( g ,w),fffa,a,) = ^^-)ln^(x9lWl), (3) H{q°,u; = - £ OK*"'**)la«•{««-»).
'ie0
"to vl(x^)=vl(x^) /
£ q»(*»»*) ; L 92€Q 9260 ui2 G f i
(4) A«° o (x^ w ?) = min max A«(x A ^ z 9 91 i"u""))q£Q u>€S"2
qiEQ wi€fl
In the version AjC2 these problems are formulated provided that the set Q consists of the unique element q corresponding to the specified normal ization u[f(x)] and for the version BjC2 utilization of the non-normalized f(x) means that Q also consists of the unique element q corresponding here to the identical normalization u[f(x)] = f(x) with x £ X. In the general case the convolution criteria are dependent on the priority incorporation procedure w, in other words, vw[Jrq(x)} — v{Tq(x)\w] with x £ X. Suppose we have the specified set W of the w priority incorporation procedures, then it is possible to determine the cases A3, B3, C3 as before and formulate the problems of choice for each of the twenty-seven versions, at the same time we can consider vw[J-q(x)) for the fixed normalizations of u, the priority incorporation of w and the convolution of v just as different convolution procedures with different w £ W, and then the cases A3, B3, C3 go over into A2, B2, C2, respectively. Consider the multiobjective decision problems under uncertainty.
360
Systems Optimization
Methodology
In the case of the unique estimate functional in decision making under uncertainty there arises the multiobjective decision problems determined by specifying the set of criteria ks = {t/],..., i/**} of decision making in various informational situations 7S of opening the uncertainty. Within the limits of the fixed informational situation Is, we can specify several estimate functional {/jj-jLi which can form a multiobjective indicator both for the specified criterion i/s 6 ks and for a set of criteria from ks. The problems constituting special cases of possible versions of formulating multiobjective decision problems under uncertainty are given below. Assume that the management unit M has L(L > I) situations of de-
cision making {X,QJ1}, {X,QJ2}, ..., {X,Q,L} differing from one another by the estimate functional, and it is required to determine one multiobjective decision for all L situations of decision making under uncertainty of selecting the state Qj € Q by the environment. The above first multiobjective decision problem under uncertainty arises when it is necessary to incorporate several indicators say, expenditures and efficiency. The multiobjective situation {X, /} of decision making in the present problem will have the estimate functionals f ( x \ Q ) depending on Q, thus we have f(xk\Q) = {f'(xk\Qj)}"f=l with xk € X. Employing the principal elements {u,v, w} of multiobjective decision making enables one to consider the situation of decision making under uncertainty with one scalar estimate functional. In the obtained situation of decision making we determine the informational situation Is, s £ {!,..., S] and choose a decision criterion. The subsequent process of decision making is carried out according to the criterion for the informational situation obtained. Describe the second multiobjective decision problem under uncertainty.
Assume that the unit M has L situations of decision making {X,Q,/1}, {X, Q, f 2 } , . . . , {X, Q, / L } differing from one another by the estimate functional. Assume that for all L decision situations the same information situation Is prevails and the unit M has chosen a decision criterion. Having applied the decision criterion in each of the L situations, we obtain the multiobjective decision situation. We next choose the principal factors of multiobjective decision making {u,v,w}, obtain the optimal solution z° € X or using now in the multiobjective decision situation the maximal efficiency principle we get the optimal solution x° £ X. Formulate the third decision problem under uncertainty. Assume that the unit M in the decision situation {X, Q, f} has the information situation
Problems of Multiobjective Optimization Under . . .
361
/s. The given information situation Is is correlated with the set of decision criteria ks = { i > * , . . . , i/**} from which the management unit M determines not one criterion but the subset k°s C ks. Applying each of the decision criteria vs € k°s to the present situation of decision making we obtain for each decision x° the vector of the estimate functional values and, consequently, the multiobjective decision situation in which a decision is chosen in the same way as had been done in the above problems. Assume that in the fourth multiobjective decision problem the unit M has the decision situation {X, Q,f} and let the set of the information situations / be specified. For each of the information situations Is C I the management unit determines the decision criterion i/° £ fc°. This situation of decision making prevails when the medium changes its state after the management unit has made a decision. Applying each of the chosen criteria to the decision situation {X, Q , f } we get the multiobjective decision situation. The last example of the decision problem under uncertainty constitutes a combination of several problems given above. Suppose we have a combination of problems of the second and third types where the management unit M has L situations of making decisions { X , Q, /' }/Lj pertaining to one information situation Is identified with the set of decision criteria ks from which the management unit M determines a subset of criteria. Solving, for each of the L situations, the multiobjective decision problem of the third type we obtain the multiobjective decision problem of the second type. The problems of optimality and improvability in the multiobjective indicator components have been considered above for the case of Paretooptimality. Here we consider the necessary and sufficient conditions of improvability in the multiobjective decision problems. Suppose we have the multiobjective decision situation {X, /}. We call the solution x 6 X = {xi,..., xmx } unimprovable (Pareto-optimal) if there is no solution x' € X
such that fl(x') > fl(x) for any / = 1,..., L, and if only one of the inequalities is strict. Hence, the solution x € X is termed improvable (on the set X by the multiobjective indicator) if there exists the solution x' € X such that f'(x') > f'(x) for any / = !,...,£/, and if only one of the inequalities is strict. The Pareto set is independent of normalization and monotone transformations when incorporating priority and choosing convolution that is why employment of the Pareto-optimal multiobjective solutions, particularly
362
Systems
Optimization
Methodology
under uncertainty, is the most promising. Moreover, the Pareto-optimality is one of the axioms (of natural conditions) of determining the principle of choice by the multiobjective indicator. Let us formulate the necessary and sufficient conditions for improvability of multiobjective decisions in the situations {X, / } . Theorem 41.1: The solution x° € X is improvable in the multiobjective decision situation {X, / } if and only if there exists the //-dimensional vector 7 = (*?i> • •. )7t) such that the inequalities cc , (; (77)),, / 'V( ^) °<) <
/o //'°(z°)
(41.1)
hold for all / = 1 , . . . , L and if only for one lQ € { 1 , . . . , L} where there is c/(l) -7i, ci(f) = = c(7) 4.1) -It,
(41.2)
f / ' ( i ) ++77;]/]. c(7) = max min [/'(*) x£A i/ = l,...,L x£X !,...,L
Theorem 41.2: The solution x° € X is improvable in the multiobjective decision situation {X, / } if the inequalities
/V) / V ) <*(/). <*(/). 1 1 == 11
l
£? fl0f (x°)
h € { ! , . . . ,L} k€{l,...,L]
(41.3) (41.3)
hold where the constants c/(/) are defined as (I = 1 , . . . , L) dU) ,(/) C
= mmf m ml/(x) ' ( x ) + cc((//)) ff m aaxx//' '((xx)) - mmin i n / 'fl((x)], a;)], IGA
LIGA
I£A
(41.4)
J
c ( / ) = max[ '(i) , maxf min /P(x)l xI ££ AA L L (£= =1 1,...,L ,...,L
(41.5)
JJ
here / ' (x) denotes the expression of the Zth component of the multiobjective indicator f(x) normalized as follows: fl(x)=
l \f (x)-minfl(y)} f'(y)\ \ f'(x)-mm I
y£X
/I \max/'fo) f max/'fo) - mm mm
J /
L y£X I/€A
f(yj\. f\y)\.
J/SA y£X
J
In this case, note that the magnitudes c(f) and c(/) of the form (41.2) and (41.5) can be obtained (for the given L-dimensional vector 7) by solving the extremal problems: c(7) a1x { A | [ / ' ( z )++ 7,] , c(l) = = A0(7) = m max{A|[/'(z) >AA, 7(]> xeR xeR1
J/ = 1 , . . . ,,L1 } ,,
(41.6)
xgX
a x {{A| A | f\x) / ' ( £ ) >> A A, , /I== 11,. , . . ..,.L, £}}. c ( / ) = A00 = o m max 1x€X 6X
(41.7)
Problems of Multiobjective
Optimization
Under . . .
363
T h e o r e m 41.3: There is no solution a;0 G A \ A such that for all I = 1 , . . . , L the inequalities (41.8) / V ) > c£,(/) ,(/) are legitimate where the constants c ; ( / ) take the form (41.4), and via X we denote the set of solutions x G A of the maximin problem (41.5), i.e., X = Ix € X : min fl{x) — max I
/ = !,...,L 1,...,L
min /'(x)M-.
iGX L/=1,...,L
J J)
The result of theorem (41.3) supplements the sufficient conditions of (41.3) for improvability of the solution a:0 G X in the situation {X, / } . Corollary 41.1: The set X C X of the solutions x G X of the max imin problem (41.5) or problem (41.7) coincides with the set X\0 — {x G X : fl(x) > \Q,1 = 1 , . . . , L} of solutions from X satisfying the maximal efficiency principle in the multiobjective decision situation {X, / } . Theorem 41.4: Any solution x € X C A" cannot be improved in the multiobjective decision situation { A , / } . Theorem 41.5: The vectors 7 = (71, • • , 7 L ) , for which the inequalities / ' ( * ) < cc,( ( (7), 7 ),
//''°°<
(41.9)
hold for all / G { ! , . . , , £ } and a particular Z0 G { 1 , . . . , L } with some solution x G A , satisfy the inequalities 7 ^2 --7 7/, / , < mmax a x /f/hl ( (x) x ) - -m mm i n / ' f2h( a{x) 7/ ;) I£A
(41.10)
IEA
for all 21}I2 G { 1 , . . . , L } . Corollary 41.2: If we assume that components of the multiobjective esti mate functional / are normalized in the natural way, then the L-dimensional vectors 7 — ( 7 I , - . , 7 L ) G G satisfying inequalities (41.10) belong to the set G(f) of the form G(f) = {7 G RL : \jh - lh\ < l},h,h £ { ! , - • • ,L}. Theorem 41.6: The solution x° G A can be improved in the situation {A, / } if and only if there exists the L-dimensional vector 7 G G for which inequalities (41.1) are fulfilled. Theorem 41.7: The solution a;0 G A cannot be improved (Pareto-optimal) in the multiobjective decision situation only when there exists the vector
364
Systems
Optimization
Methodology
7 6 G such that [f'{x°) + 7z] 7;] = max min [[/'(x) / ' ( * ) + 7/] • min [/'(x°)
1 1,.. ,t / = =1,...,L
x£X l =1,...,L \,...,L iG X 1=
(41.11)
Corollary 41.3: Condition (41.11) of theorem (41.7) is sufficient for con cave / ' (x) with at € X, I — 1 , . . . , £, i.e. the set of Pareto-optimals solutions in the situation {X, / } coincides in this case with the set X° = {x € X\ 37 € G : (41.11)}, which determines characterization of the set of Pareto-optimal solutions differing from the tradeoff domain. Theorem 41.8: For the solution x° € X to be improvable (Paretooptimal) in the situation {X, / } , it is necessary and sufficient that the following inequalities are fulfilled (inconsistent) f'(x0)
< max{max[
min (/''(at) + 7*<)j 7c) - 7/} •
The proof of the above and subsequent statements is given in the au thor's work.8 In the multiobjective decision situation {X, / } under uncertainty it is possible to obtain the necessary and sufficient conditions for improvability of the solutions x° E X, which is achieved in the same way via concretization of multiobjective information situations. Consider some problems of multiobjective decision stability. Suppose we have the multiobjective decision situation {X, / } in which X — {xi,... ,xmx}, f(x) = {/ i (x)}^ =1 , here the components of f(x) are specified in a positive ingredient. We say that the solution x° G X is Pareto-optimal on X in that there is no as* £ X for which the inequalities f(x*) > fl{x°), VZ € { 1 , . . . , L } are fulfilled and if only one inequality is strict. Let the set of Pareto-optimal solutions for the multiobjective indicator / be denoted via M(f). The stability of multiobjective decisions is understood as the property to remain optimal with small changes in the specification of the relation of natural order Ndom(f) used in the Paretooptimality principle. Define the following estimation of stability of the solution x° on the set X : e(x°) = min ||/(x 0 ) - f(x)\\N*"x , here || • |AT™X
Problems of Multiobjective
can be defined, say, with n £ Nmx ||n|| = \\n\\
max
Ki<mx
Optimization
Under . . .
365
in the following way: max[0,n t ],
||n|| = YJ max[0,rii], = ii i; =
<5(i°) 8(x°) =
min
x€X\M(f) x€X\M(f)
||x° a;||jvmmAx \\x° — a:||jv
Denote by y the relation of order Ndom(f). 8 > 0 define the sets M£e(f)
For arbitrary numbers e > 0,
= {x £E X\ 3x' € X : /(*») X /(*)} , £■
M«(/) = {x €G X\ X\ 3x' 3x' ££ XX ::f(x') f(x') yy f(x)} /(*)}, , 6
here by >-, >- we denote the relations of order N(f) C NL determined by s * the following conditions f(x') y f(x) if and only if there exists a £ iV , a > 0, ||at||#s <
£ sucn
that f(x') y f(x) — a; f(x') y f(x) if and only if 6
>6,f(x')yf(x). We call the sets Me(f), M$(f) respectively the regions of e and 6 stability, and the magnitudes e and 6 the radii of multiobjective decision stability. Thus we have: \\X'-X\\N>"X
Theorem 41.9: The conditions Mc(f) C M(f) C Ms(f) are fulfilled. The absolute stability domain is identified as Ma(f) = {x £ X\ 3x' £ X : f(x') y / ( * ) } , here / ( i ' ) ► / ( i ) with a = (e,c5) £ i? 2 if and only if [/(*») J- /(*)] A {[ | | i ' - x\\Nmx a< 6} V [/(*') y /(»)]} and M a ( / ) C M ( / ) . Suppose we have the following sets with e > 0, 6 > 0: M;U)
= {x £ X\3x'
£ X : f(x') y f(x)} , £e
M* ( *1')) yy f(x)} M ;6((f) /) = = {x {z £G X\ X\ 3x' 3x' ££ XX :: /fix f{x)} ,, 6
L
here /(a;') y f(x) if and only if for any a G N , a > 0, ||a|| w i. = 1 the conditions f(x') y f(x) - a are fulfilled, similarly f(x) if f(x') y f(x") 8
e
for any x" G X such that [|s - x"\\Nn,x
Theorem 41.10: Fulfill the following relations:
M;(/)CM(/)CM;(/).
y f(x) if and only 6
< 6.
366
Systems
Optimization
Methodology
Define with a = (e,6) the set M*a(f) =
{xeX\3x'eX:f(x')yf(x)}, a
here f(x') y f(x) if and only if for any x" € X such that ||£-2"||N m x < 6 a * the condition f(x") >- f(x) holds. Suppose we have the set S(f) = {x € X\3-y £ NL : f{x) -< 5(7)}, here c(l) = (ci{l),-..,cL{f)), 0(7) = c ( 7 ) - 7 ; , c ( 7 ) = max min f/'(x)+7,]. T h e o r e m 41.11: There exists M(/) =
S(f).
Define Ss(/) X | 37 3 7 GG NNLL ::/(re) /(re) -<-
5;(/) = = {* { xG€ X e ^^: / ; :( /x;)(^*cH( 7c(7)}, )}, s;(/) x|| 3377 e £
L
5 ; ( / ) = { x eGXX| |3377€G^ ^ : / ( x ) ^ c : ( 7 ) } , where Cf(7) = ce{i) =
max
[/'(z') + 7/], 7,], min [/*<«*)
| | l - xl ' | | N m Xx
/?(») =
c*(7) = = 0(7) + e , c*(f)
><5
max x, ex ll*-» ll*-» ,fl*~* fl*~* <* <*
f(x').
T h e o r e m 41.12: The following conditions are satisfied: 5^(/) =
s*s(f) = M;(/),
S.(/)
= M£(/), s;(f) = M*(/).
Ms{f),
Consider the set B(f) of the form B(f) = {x G JC| 37 € NL : f(x) y c(-y)}. We have S(f) C B(f). Similarly, it is possible to define the appro priate sets Be(f), B*(f), etc., and also the set (/3 > 0) S0fi(f) {t) = {x G X\ 3 7 G € NL : /(«) xX ^ fi»( ( 7 )7})},, c/3(7) =
max y g wLL -i'£N
ll7-7'll«".x3
max min f fl(x') L ii'exi
+ ■y',]. ""
Problems of Multiobjective
Optimization
Under , . .
367
The above concepts of stability make it possible to extend the formu lation of the solutions optimal by the principle of choice on the basis of generalizing the natural order Ndom and the concepts of improvability of solutions by the multiobjective indicator so that they reflect different pro perties of sensibility, stability invariance, tolerance, vagueness, etc. of the relations of order and appropriate optimal multiobjective solutions. §42. MULTIOBJECTIVE OPTIMIZATION PROBLEMS FOR D Y N A M I C CONTROL SYSTEMS The vector optimization problem for dynamic control systems with flex ible extremities is referred to as the problem of the form: obtain the controls
x0< ! Xo < xX<<XXi
(41.12)
otherwise y = y0 with x = x 0 , y = 2/1 ^vith x = x%\ M\ is a set of elements {x0,xi,yo,3/i>¥>} G R1 x R1 x Rn x Rn x L*\z0,Xi], such that a0 < XQ < b0, ai < x < h, tn(sco»»itIto.Kl) = 0 ( 1 = l,...,fc),
/o(y(&|v>),¥i(aj),s)da;, /od/C^lv).^*).*)^!
J x0
Vi(x = gi(x Vi(x00,xi,yo,yi,v) , xi, 2/0,2/1, ¥>) = <7t(a=o, xi,j/ 0,xi,y 0,yi)+ 0 ,J/i)+ /
J x0
/i(2/(s|¥>)>
368
Systems Optimization Methodology
Let there be given the numbers of ao, 60, a-i, b\ (bo < &i) r-dimensional
vector-functions ^(x),
2/012/i), 9i(xo,x\,yo,y\) continuously differentiable with respect to XQ € [a 0 ,6o], x\ £ [ai,6i], j/o 6 Rn, y\ € Rn From the mathematical point of view it might be good to improve the vector optimization problem formulation, indicate the rule in the sense of how one should understand the solution of the multiobjective optimization problem for the dynamic control system with flexible extremities when the constraints have the form of equalities and inequalities (the principle of choosing an optimal control according to the vector criterion VQ). Consider natural conditions or axioms for the principle of choice in the dynamic control system with flexible extremities. We call the order (or preference) the binary relation TV between two
elements {zj,z},^,7/J V}, {x%,x^,y%,y%,
(1) {xQ,Zi,yQ,yi,v}N{x0,x1,ygfyi,tp} for any {XQ,II, 3/0,3/1, ¥>} 6 MI n M-2 (reflexivity);
(2) if {xlxlylyl,v1}N{xlxlylyl
I/O' y?'^3}' then {^o^i'yo'l/i^1}-^^^1!'^^!'^3} holds for any {lo.^i.J/oiJ/i.V1}' {zo>*i,2/o,2/i.>2}- {^o.^i-J/o.J/i-V3} € M!nM 2 (transitivity);
(3) either {xlx\^lyl^}N{xlxl 1
1
ylyl^},0T{xlxlylyl^}x 1 2 €
N{x 0, x\, yj, yj.ip } for any {zj,ij,7/i, 7/J,^ }, {^O'^i'^O'J/i'V } Mj n M2 (asymmetry);
(4) if {xlxlyLylvWxl^ylylv*}, {xg.xj.^.y?,^}^^, xj, 2/o. 2/i.V1}, then the elements {xl0,x\,y^,yl,
Problems of Multiobjective Optimization Under . . .
369
2/0'2/i-V1) < V"0(:rg,2;?,j/g,2/2,(/j2)forany {x^x\,y^,ylf1}, {xl,x\,y%,y\, V?2} € MI nM 2 .
Each of the components VQ (s = 1,..., N) of the vector criterion establishes the natural order on the elements {x0,xi,y0,yi,(p} € MI n M2 we denote it by Ns. Assume that for the elements { x 0 , x i , y 0 , y i , < p } € MI n M2 and the vector criterion V0 = {V^,..., V0N} the relation of order N(V£,..., V0N) is defined, then the vector optimization problem for a dynamic system is the problem of deriving such {a;§, x\, y$, y°,
(2) the condition of independence when choosing the elements {XQ,XI, J/Q,
3/1, if] € MI n M2, in other words, if for any two vector criteria there exists *V0 = {*V 0 1 ,...,*V/}, °Vb = {°VJ,...,°V0N} } the elements {x^x^y^yl,?1}, {*o,zi,3/o,3/i,¥>2} € MiHM 2 are such that *V0(x%, x\, yl yl ^) = °V0(xl xl yl y{,
370
Systems
Optimization
Methodology
order N(V0l ,. .., VQN) for these elements, i.e. 2 2 2 3s e {l,... ,N) : {xl,x\,yl,y\,^}N{xl,x\,yl,y\,^}N } s{xl,x\,yl,y\ s{x ,x\,y ,yVV} 1 1 1 1 NN 2 2 ^{x10,x ,v1}N(V0\...,V ){x2,xlylyl
(5) the weakened "Pareto principle," thus we have (Ni x N2 x • ■ ■ x N^) C N(V£,...,VQN), i.e. if the element { i g , ! ? , ^ 0 , ^ , ^ 0 } £ M , n M 2 is optimal according to each of the special criteria V0s(s = 1, ...,JV), then {XQ , 1J, j/o' 2/i > V° } is optimal according to t he principle of choosing N{V0\..., V ^ ) denned for the vector criterion V0 = (V01,..., V0N). As applied to the formulated natural conditions we have the problem of the existence of the relations of order N(VJ,..., V0N) satisfying these con ditions (the investigations into this problem resulted in the Arrow paradox of nonexistence of the order relation satisfying these axioms). Thus the efforts to change the formulation of these conditions so that there might exist the relation of order N(VQ, . . . , V0N) satisfying even new requirements are quite legitimate. Assume that V0 = (V0\ . . . , V0N), 'V0 = {'V0\ ... ,'V0N) are two vec tor optimization criteria for the dynamic control system, and Ni,... ,NN\ N[,..., N'N are the natural order relations determined by the special crite ria VQ1 , . . . , V0N;' VQ1 , . . . , ' VQN respectively. In this case % = {'V0\..., 'V0N} are referred to as the improvement of Vo = { VQ1 , . . . , VJ> } (and Vo the dete rioration of 'VQ respectively), if from fulfillment of the condition {xg, i j , ^ , S i . ^ K M x ■■■ x NN){xl,x\,yl,y\,ip2} for the elements {xl,x\,yl,y{,
ip1},
(a) {xl xl yl y\, ^}N(V0\.. .,V0N){x2, x2, y2, y2, V2}, % = ('Vo1, ..., 'V0N) for improvement VQ = {F 0 V . -,V0N}, then {xj,xj, yl0,y\,
Problems of Multiobjective Optimization Under . . .
371
(b) {xl x\, yl yl ^}N(V0l...,V0N){xl if, t/ 02, yl ^}, 'V0 = N N {'VJ,...,'V} with deterioration V0 = {Vj,..., V0N}, then 0 l N {xl i{, yl ylv }N('V 0l...,'V0 ){xl z2, yl, yl ?2} the line above the order relation denotes negation of the relation; (4') the condition of the free principle of choice which resides in the fact that for any elements {z^zj.y^y},^ 1 }, {z^z 2 ^,?/ 2 ,^ 2 } e MI fl A/2 differing from one another there exist the vector criteria VQ = {VQl...,V0N},, % = {'V0l...,'V0N}} such that {xl x\, yl y\, } N(V0l ..., V»){xl xl yl yl
372
Systems
Optimization
Methodology
Definition 42.1: The element {x%,x1,y$,yt,
(s = l , ......,,AA^0)-.
Definition 42.4: The element {xg,a:?,t/g,z/°, v?0} e Mx n M 2 is KuhnTucker efficient if there is no element {xo,xi,yo,yi,
Problems
of Multiobjective
Optimization
Under
.. .
373
| [ g ^ a dIIoV o ^ (0 ^'(4,x;,i,S,»?, , x ; , % ^ 7 / ° , ( pv00))]xo ] x • 0 + [grad oy ro0s8((xg,x; /^ ( pvj00)]x |[grad [gradsxoV )]xi1 a :°,x?, % ^ )2j/?, ) 2/g [ g ryoV a d0°(x ^ x00^,xly x 0^0,ylv o 00)ryo , ^ , / ) ] ^ ++ [ g[ rgarda ^d x^ gV. yx ^^^y.gK. gy .Sr V f V) ]j ^r f! t +■[g™d + / VadvoV0<(x°,x^0^°,)]V(x)
S = l,...,iV
0 K*(x°,x?, ,?,/) K*(x°,x?, 2/g,0 1/? ,¥>") =
G
Mx n M2 is called
mm m i nW /( *. 8I. *^i .^s 8A. yi? V / ) ■ .
3 = 1,...,/V
Definition 42.6: The element {xg,x?,2/g,j/i,^ 0 } € Mj n M 2 is called the tradeoff one (or belonging to the tradeoff domain) if there exists Q° = (<4,...,tx%)
£NN
N
such that a° > 0{s = 1
N
Va°^(x°,x°,y V a ^ ( x ° , x 0°0 ,2,°,^°)= ,«V)= ■ s = l* 3= 1
N), £ Q° = 1, s= l
N
min ^r a^ ^' ^x. o^ .. fxc ^. iof ,i .^ v) ) •.
M\C\M2 *-~* s=l 3=1
Definition 42.7: The element {xg,x?,?/g,?/?,} £ M x n M 2 is called minimal if min
3 = 11,...,N ,...,w
V r 0 s (xg,x?,j/g,j/?,^°)== V£{x%,x\,y%,y\,v?)
min mla„V?(x 8 ,*i,Jte>Si,v)| •■ m i n [[ min„Vy(aJo,»i,Jto,Si,v)|
M i n J w 2 LLsS== Il,...,Jv MinJvf ,...,JV
Definition 42.8: The element {x%,x1,y$,y^,
s=l,...,W
Jj
£ M x n M2 is called
^VJ(x%,x1,y%,y%,
J
Definition 42.9: The element {xg,xj,7/g,y?, y 0 } 6 Mi n M 2 is called efficient from below (or minimax) if s max V 0r 0s (x8,x?,y2,g?,
s=l,...,N
min
max V 0 s (x 0 ,xi,?/o,2/i,v) ■
MinAf2 L s = l,...,iv
J
374
Systems Optimization Methodology
D e f i n i t i o n 4 2 . 1 0 : T h e element {x%, x^yfi, efficient from above (or maximin) if min
5 = 1,...,7V
V0s(x°0,x01,y0Q,y01,
max [
j / ? , ^ 0 } £ M2 n M 2 is called
min
JWjnA/2 L s = l , . . . , W
YSfa.XuVo,)/!,?)]
■
J
D e f i n i t i o n 4 2 . 1 1 : T h e element {x%,x1,y%,y%,tp0} € Mx D M 2 is called co efficient from below (or Gurvitz-efficient from below with the a indicator) if there exists 0 min V^( V^xlx\,ylyl^) + ({l-a) mai V y o0 ]' ( iz»° ,, i: !r ,^j °S, ,j j, »S ,V/ )) l| f a min l - Q ) max a ; °,x?,3/o ,2/ 1 ,/) + L s = l,...,N s = l,...,N J
=
min a min F0**(a;o,a-i,fto, yi,v) (xo.^uffchftu^) MinM2L s = l,...,N MiHM + ( 1 - Q ) max ^ ( x o V, 0xs(x ,yi,ifi)\ i ,0,xi,y 3/0, 0J/i,v) s=l,...,N S =l,...,/V
J J
Here the value a(0 < a < 1) is assumed to be specified. D e f i n i t i o n 4 2 . 1 2 : T h e element {a^, y?, 3/°, (/?, v?°} € Mx 0 M2 is called co efficient from above (or Gurvitz-efficient from above with the a indicator) if there exists [o \a L
max
ss=l,...,N =l,...,/V
=
V o0 ss ((xx°°,,x%, %° 0, ,22//??V , /) + ( l - a )
min mm
s=l,...,N
W < * g , x ° , il & » ? , / ) '
a max V V00ss(xo, (xo,ii,3/o,2/i,y) max cv 11,3/0,3/1, v ) MiC\M2\- s = = l,...,N l,...,N A^nA^L + ( 11 -- QQ) )
min min
^VQ^^O, ( 0 : 0 , 3^1,3/0,3/1, : 1 , y o , y i , v<) p )■. JJ Here the value Q ( 0 < Q < 1) is specified. SS = =\,...,N \,...,N
D e f i n i t i o n 4 2 . 1 3 : The element { a & x j . j g , j/?,
max if ff[Vb(zo,*i,3/o,*,*)] [ V 0 ( z o , * i , 2 / o , * , * ) ]-,
MinM2
where H[V0] is the uncertainty function of the values of the vector criterion V"o determined, say, when H is taken to be the Shannon entropy H[V0(x {x00,xi, ,xx,
NN 3/0, ( i o , a&i,ito,¥i, xi, 3/0,3/1,9), (*0i*l*i,2/0,2/1, 3/0, jj/i, / ! , v?)] = i#f[^^ 10 (»o, * > ) ,•. .•. ,•,FV - 2/o, 2/1. ¥>)] f)] 0 0 (*0i
N
= -^2^(x0,x1,y0,yi,
,
Problems
of Multiobjective
Optimization
Under
. . .
375
here {s = 1 , . . . ,N) N
Vo(x0ixi,yv,yit
^o( ^o(xxo,xi,yo,yi^), o,xi,yo,yi^),
= ^(aso.ari.ito.Si.vO / X (7 c r== l1
^^(^o,-?!,2/o,2/i,
13
v
min *o V0s(a;o,xi, y0,yi,
min
Mi n M2
x
..F F 1 3 ??,, ^ ( ^ 0o( , ^ o,xi,y 1 , ^ 0 0, ,yi,
~ M »,i min , V^oC^o,^1,2/0,2/1,^)1 o (^0,£1,2/0,01,V)IJ MinAf nM2 2
■
Definition 42.14: The element {x%,x\,y%,y\^} € Mi n M 2 is called op timal from below by the A-criterion if we fulfill the condition {xg,x?, 2/o°, y\> i/?0} € MAO, where Af\ denotes a set of such elements {a;o,a;i,l/o,2/i,'} € Mi n M 2 that Vjf(a;o,xllj/o>yi,
min
0<X
A,
where the functional VQ take the form as given above. Definition 42.15: The element {x%,x1,y%, y^f0} £ Ml n M2 is called optimal from above by the A-criterion if we fulfill the condition {XQ,X°, 2/o,2/i> V0} £ MAO where MA denotes a set of such elements {x 0 ,xi, 2/0,2/1,^} 6 Mi n M 2 that V0s(x0,£i,2/o,2/i, ¥>) > A, s = 1 , . . . ,7V and the magnitude A0 is the solution of the extremal problem A°0 = A
max
A.
0<X
Consider the conditions for the dynamic control system improvability according to the vector criterion. Assume that the feasible {xl,x\,yl,y{,if1} € Mi n M 2 and the vector criterion V0 = {VQ, . . . , V^} which determines the trivial relation of order N(VQ, ... ,V0N), are specified. The question arises as to whether there exists another feasible set {:ro,z?,2/o,2/i>¥>2} £ Mi n M 2 which is more
376
Systems
Optimization
Methodology
preferable than {xJ,£i,2/o,2/i,?1} € Mx n M 2 in the sense of the trivial relation of order N( Vj,..., V0N) = JVj x • • • x NN, i.e. for any s = l,...,N the inequality s 1 VQ0*(xlxlylyl
(41.13) (41.13)
holds and at least one inequality is strict, what the conditions for the exis tence of such an element {xl,xl,yl,yl,tp2} € MiflAfj are. Here we suppose that approaching an optimum of the vector indicator resides in increasing each of the components of the vector criterion VQ = {VQ, . . . , VQ}. The following theorem formulates a sufficient condition for improvability of the element {XQ,x\,yl,yl,i^} € Mi n M 2 according to the vector criterion. Theorem 41.13: The feasible element {rrj, x\,y\,y\,
(41.14) (41-14)
is fulfilled, here the constants CS(V0) are defined as follows (a = l , . . . , J V ) : C , 0 4 i ) =
VDa(x0,x1,yQlyi,¥>)
min + C(V0)0) + C(V -~
max ^ ( I O VJ{x max ^ I .0J,xi,y / O .0J,yi,(p) /I,^) L MinMj min Vv ! (ix ,ii,!/o,!/i,ip) min
,„
,
o( 0o,xi,yo,ywfi)\
0
(41.15)
where the constant C(V0) takes the form s C(V0)= max f min min K 0 s (x V0 ,a:i,J/o.3/i,v) {x0,xi,y0,yi,^) ) = max.
Mir\M L s= MinAf2 = l,...,/V l,...,N 2
(41.16)
and V'0S is a normalized value of the sth special criterion V0S. Condition (41.15) is not only the sufficient condition for improvement but also is close to the necessary condition. Thus we have Theorem 41.14: There is no feasible element {x^xl^^y^tp1}
G Mi n
M 2 for which the inequality (5 = 1 , . . . ,JV) 1 1 (xlx\,y0l,yl
(41.17) (41.17)
Problems of Multiobjective
Optimization
Under . . .
377
holds if among special criteria of the vector criterion V0 = {VQ, . . . , V0N] there are no degenerate ones. Degeneracy of the special criterion V0S of the vector criterion V0 — {VQ,..., VQ } is taken as fulfilment of the equality s V0s(x = min min V V00SS == max max V V00SS == const. const. Vo (io,ii,!/o,!/i>v)= 0,Xi,y0,yi,ifi)
MinM2 M\C\M2
Mi nn A/2 A/2 Mi
We set forth a generalization of the improvability conditions. Theorem 41.15: The feasible element {xl,x\,yl,y\,y?1} 6 Mi n M 2 can be improved according to the vector criterion Vo = {VQ, . . . , V0N} if and only if there exists the vector 7 = (71, ■ • ■ ,7JV) € RN such that the inequalities 1 1 1 VJixlxlylylv )^^), ^(xJ,xJ, 2/ J,i/ 1 , v? )
V^xlxl^lyl^XC.Cr) <(xJ, a; J,^, 2/l 1 , V 1 )
(41.18)
hold for any s € { 1 , . . . ,N] and for at least one a € { 1 , . . . , N}, where =C C((1)-j = l , - - l,...,N), 1) = Css((7) 7 )s(s -7s(s = -,Ar), C(j) = C(7) = max { min [V0s(x0,xi,y0,yi,ip) JUinA/2 ls=l,...,JV
+ 7,] I.\. >J
If the feasible element from Mx n M 2 is specified, the problem arises with respect to determining in RN the range of values of the vector 7 for which inequalities (41.18) are fulfilled. Theorem 41.16: Any vector 7 G RN for which (41.18) are fulfilled with some feasible elements {x^xl^^y^f1} € Mi n M 2 satisfies inequalities for all s, p£ {1,...,N}: a {x00,xi,y ,xi,yo,yi,(p)max V0s(x 0,yi,
AfinM2
min V' V00p''(:ro,xi,yo,yi,¥>) (x0,a;i,?/o,i/i,(y9) >> 77pP - 77ss •• (41.19) (41.19) min
MinM2 MiOM 2
We have the following equalities from the definition of the normalized value VQS with s = 1,.. ■ ,N s min V ^(xo,xi,y 0 (x0,xi,yo,y 1,
MinM MinM 22
s max. V -= 1 V^(i ,a;i,3/o,!/i.v) o (x0 ,x 1,yo,yi,f)
MinM MinM 22
378
Systems Optimization Methodology
then the range of values of the vector 7 6 NN determined by inequalities (41.19) coincides with the set G of the form
G={^eNN :7s-7,< 1, VS, pe {!,...,N} (41.20) with N = 2 we have G = {(71,72) e TV 2 : ^ -7 2 | < 1}. Theorem 41.17: If inequalities (41.18) hold for {zj.z},^,^ 1 ,^ 1 } € MI n M2 and a particular vector 71 e A r/v , then they also hold for 7* = 71 - 6 = (7! - <5,72 - < 5 , . . . , 7]^ - <5) with anyfinite5 6 Rl • The set G (41.20) can be restricted, say, to the set G' = {7 6 TV^ : 7! = 0, 7S - 7P < 1, Vs, p e {!,...,JV}, in particular, with N = 2, G' = { ( 7 i , 7 2 ) e R 2 : 7 i = 0 , | 7 2 | < l } . Corollary 42.1: (to theorem (41.15)). The feasible element {zj,:rj,2/o, y}, i/?1} 6 MI n M2 can be improved according to the vector criterion V0 = {Vj,...,0NV } if and only if such that there is a vector 7 € G. Theorem 41.18: The feasible element {ij,2/0^1.2/i> V 1 } € M! n M2 cannot be improved (Pareto-optimal) according to the vector criterion V0 = {VQ ,..., VQ*} if and only if there is no vector 7 = ( 7 1 , . . . , 7 n ) £ G such that inequalities (41.18) are consistent. Note, that application of the maximal efficiency principle to solve the multiobjective optimization problems for dynamic control systems enables one to formulate the relation of order N\ satisfying the axioms of choice and leads to an extremal problem (A-problem) of the form for which the apparatus has been adequately developed for obtaining optimality conditions and numerical techniques of solution, in other words, the maximal efficiency principle looks rather promising for applications when solving the multiobjective optimization problems of dynamic control systems.
Chapter 12 METHODOLOGY OF VECTOR OPTIMIZATION
§43. OPTIMIZATION METHODOLOGY FOR HIERARCHICAL SEQUENCE OF QUALITY CRITERIA We formulate, in a simplified form, the vector optimization problem as follows: assume that system behavior is characterized by the n-dimensional vector x = {xi,... ,xn} 6 X C En and estimated by a fc-dimensional vector-function I ( x ) = |/i(x),..., Ik(%)} the components of which constitute specified real functions of a variable x, where it is required to find a point of x° £ X optimizing, in a sense, values of functions / i ( x ) , . . . , I k ( x ) . Many of the studies on vector optimization methodology can be grouped according to the following directions: - optimization methodology of hierarchical sequence of quality criteria; - methodology of determining the sets of unimprovable points; - methodology of determining a tradeoff decision. We consider all these approaches in succession. Optimization methodology for hierarchical sequence of quality criteria is based on introducing an order relation on a set of criteria. Assume that an order is set such that /i (x) > / 2 (x) > • • • > //b(z), the solution of x° € X is defined as satisfying the relations:
379
380
Systems
Optimization
h(x°)=
min
Methodology
!,(*),
iGAiCAo
I3(x°)= I7kfc{x°) (2:0) =
min
/,(«),
i€A2CAi
min I€AI.-ICAI,-2
J fcI(x), k(x),
here the sets X; C Xi_i (z = 1,2,..., fe- 1) are defined as Xx = {x : / i ( i ) = min 7i(x)}. When solving most of the practical problems, employment of that kind of approach proves to be ineffective since optimizing even by the first cri terion leads to a unique optimal solution (if any) and everything results in optimizing only by the first criterion. The problem arises with respect to determining relative importance of quality criteria in vector problems of managerial decision optimization. We consider the problem in its general statement. In vector optimization problems, "utility," "quality," "efficiency," "value" and other features of objects are estimated by the criteria k\, fc2,. • •, km (in > 2) generally identified as the function mapping a set of objects Q in a subset Xt (containing no less than two points) of a numerical line Ri which is known as the zth criterion scale, and its elements are scale estimates. A set of criteria fc, forms the vector criterion fc = (fci,... ,fcm) mapping the set Q in the set \ = Xi x Xi x ■ • ■ x Xm of vectors, the components of which are scale estimates. For analyzing a particular prob lem and obtaining the necessary information to solve it, it is convenient to manipulate vectors from \ which do not necessarily correspond to real objects from Q but which can be regarded as characteristics of some hypo thetical objects; such vectors are commonly identified as vector estimates. In this case, some combinations of scale estimates may prove to be unfea sible because the vectors involving these estimates will be the meaningless sets of numbers. Hence, in general, inclusion of X C \ is strict. Based on the information obtained on the problem, in the set of vector estimates A' we construct the binary relations of preference and indifference which are then, along with supplementary hypotheses and considerations (in particu lar, the maximal guaranteed result principle employed when indeterminate factors are available), used for constructing preference relations on a set of strategies U and isolating an optimal strategy. Furthermore the relations of preference, indifference, nonstrict preference and incomparability will be
Methodology of Vector Optimization
381
respectively labelled by P, I, R, N which, if need be, will be provided with appropriate super- and subscripts. According to what has been previously said in our previous discussion, for any two vector estimates x and y only one of the following cases is possible: (a) xPy x is preferable to y; (b) yPx; (c) xly x and y are the same by preference, i.e. it does not matter which of the two vector estimates is selected as the best; (d) xNy x and y cannot be compared by preference. xRy should imply that either xPy or xly is true, in other words, there exists R — P U I. Relations P, / and N are recovered by R (we can say that the relation R generates P, I and N): xly when simultaneously xRy and yRx; when xPy, xRy but yRx is not true; xNy when neither xRy nor yRx is true. Assume that criteria are independent by preference and, for each of them, the large values are preferred over the small, then the vector estimate is preferred over the vector estimate (o;||j/i) obtained from x by substituting Xi by the scale estimate yi with any i € M — {!,..., m}, yr < xt. It
is natural to introduce into X, and formally also into /?/»>, the relation of nonstrict preference R° defined as follows: xR°y when there exist the inequalities x^>y^, i — I,... ,m. (43.1) Evidently, the relation R° generates the relations 7°, P° and JV° It is clear that xP°y when if only one of inequalities (43.1) is strict, and 7° is the relation of vector equality. In the analysis of vector optimization problems the relations introduced are rather useful, but they do not allow one to solve concrete problems in full since many vector estimates prove to be incomparable. A need arises for extending these relations, and it can be done only when obtaining additional information. We will consider constructing the relations of preference and indifference using information on the importance of criteria as the base. Information on the relative importance of criteria is given by a totality of letters, figures and other signs, i.e. it can be represented by the set fi, we will also employ the set fJ° = ft U {0}. Each message w € ft enables one to compare the vector estimates of special types by preference, in other words, to establish in X the relations of strict preference Pu and indifference /"', or else only one of such relations, R" is taken as a combination of these relations. Since, for comparing two vector estimates it is possible to employ the relation R" for any LJ € fi, as well as 7?°, then the information 0 determines in X the binary relation of nonstrict preference R^, strict
382 Systems Optimization Methodology preference P^ and indifference /P, thus (43.2) The relations introduced are beneficial for solving vector optimization problems. The relations F,n, / + n , .R", generally speaking, are not transitive. If it is desired to have the transitive relations of preference and indifference P n , 7 n , J? n in the problem, then they can be obtained by extending the relations. Assume that for the vector estimates x and y there are such s of vector
estimates zk and s + 1 of symbols uik € fi° that the statements xR" zl, zlR"*z2, ..., zsR"°+ly^>xR"lzlR^ . ..zsR"'+1y are valid (43.2). If is not transitive, then xR^y can be untrue.
Supplement R^ with the pairs of estimates (x,y) for which there exist chains of the type (43.2). Extension of binary relations in this fashion is termed a transitive closure (denote it by TTCl] and the extended relations the transitive closures of initials. Using information fJ on the importance of criteria as the base, we then construct a reflexive and transitive relation of nonstrict preference — a quasi-order Rn = TrClR^. Since for vector estimates x and y there exists a chain of the type (43.2) in which all R are /, then these estimates are to be taken as the same by preference. Thus, the relation of indifference (^-equivalence) / n = TrClI^ is defined in X. Since for vector estimates x and y there exists a chain of the type (43.2) in which even one R is P, then x is assumed to be preferable to y. By this means the relation of strict preference Pn is determined. Hence Rn = / n u Pn and Pn offers the property of generalized transitivity by / n from xP^y and yl^z, as well as from xlny and yP^z we have xPnz. We consider the definition of ordering by the importance of homogeneous criteria. Criteria ktl,..., fc!n are termed homogeneous if they have a common scale Xtl = • • • = Xin and a set X C Rlm is symmetrical in reference to coordinate axes X^,..., Xln, in other words, if for any vector estimate » = (*!,..., xm) it can be stated that the vector obtained from x by any rearrangement of coordinates i n , . . . , x,n proves to be a vector estimate, i.e. belongs to X. In general, the idea of symmetry is used and underlies any concept of equality, equivalence, etc. and conversely any idea of preference, superiority, etc. is based on negation of symmetry defined in one form or other.
Methodology of Vector Optimization
383
Suppose that we have two homogeneous criteria kr and kt and the vector estimates x and xTt are symmetrical and obtained from x by rearranging its components xr and xt. If two vector estimates x and xri are regarded as equally preferable, then the criteria kr and kt should be regarded as equivalent, equally important. Definition 43.1: Criteria kr and kt are equivalent (denoted by rSt) when any two vector estimates x and xTt are equal by preference. Assume that the stated circumstance represents the message wi, then we say that the message u>i establishes the relation of indifference IWI between vector estimates x and y = xrt : xlwiy.11 The best of any two estimates is thought to be that (x or xrt) in which the rth coordinate is larger, in such an event, the symmetry of the rth and xt is
preferable to y = xTi(xP^y}. Consider the problem of comparing a combination of criteria by importance. Assume that we have two non-intersecting sets of criteria (n > 1, / > 1) {fcr i , . . . , f c r j , { f c t l , . . . , M - ( 4 3 . 3 ) Suppose that all criteria are homogeneous, we have a homogeneous set n + I of criteria. In homogeneous sets for ordering criteria by importance we have to compare only vector estimates, thus, x, y = ;r{ r i>---> r ™H*i'--->*'}; xTl = ••• = xTn;
xtl=--- = xt,,
(43.4)
here z'Cri.i—Jr»M*ii—«*i} js the vector obtained from x by replacing each of n coordinates xTt by xtl and each of / coordinates xtj by xri. Definition 43.3: Homogeneous sets of criteria of the form (43.3) are equivalent to u>3 = {TI, . . . ,rn}S{ti,... ,ti} when any two vector estimates (43.4)
are the same by preference (xl"3y). Definition 43.4: From two homogeneous sets of criteria (43.3) the first is more important than the second w4 = {TI,..., rn}B{t\,..., tt} when from
384
Systems Optimization Methodology
any two vector estimates (43.4) with xri > Xtl the first is preferred over the second (xP^y). Practical problems make good use of transformations of the initial inhomogeneous criteria fc, into the homogeneous ki criteria in the following way:
here b1 are some standard values of fo; a; are minimum feasible (possible) values of A;,. By no means do we always manage to transform inhomogeneous criteria into homogeneous and then order them by importance. It is desirable to determine the relative importance of inhomogeneous criteria for which we need to correlate the influence on a common quality of the objects or phenomena under study, where changes in values of these criteria exists in the scales corresponding to the criteria on hand. It is called to compare by utility the differences of criteria values. Such comparisons take place only when the differences of criteria values are meaningful. The criteria of that kind are assumed to be identified as those possessing solid scales. The criteria with solid scales may be exemplified by the criteria with an absolute scale, a scale of relations (the unit of measurement can be fixed), a scale of intervals (a scale can be specified), the criterion with a rating scale where rates are prescribed according to established standards or assigned to initial non-numerical (clearly stated) estimates. V. V. Podinovsky11 investigated the problems of estimating relative importance of criteria or their combinations (w-tuples). Furthermore, he studied not only superiority by importance but determines how much the superiority is. §44. OPTIMIZATION OF HIERARCHICAL SEQUENCE OF QUALITY CRITERIA Vector optimization problems were initially encountered in the works by Pareto.24 The works10'19'20'22 touched upon the issues of multicriterion problem, though the problem did not obtain further development.15 A deeper insight into the vector optimization problem was gained by L. Zadeh,26 who was the first to call the experts' attention to the fact that in the majority of practical cases, when developing complex systems, they tried to obtain the result possessing numerous optimal characteris-
Methodology of Vector Optimization
385
tics. Zadeh showed that exact optimization of the vector functional is generally unattainable. This means that if, by selecting the control, we manage to optimize some scalar functional, then, almost without exception, no possibility is left, with the same solution, for optimization of another scalar functional. The problem of vector optimization both in static and in dynamic systems was studied by A. Klinger, V. Nelson, M. Chang, Da Cuhn and E. Pollack, Don Hak Chuang, F. Waltz, and others.16 In the last few years the problem of vector optimization has been extensively developed since numerous problems of decision making in the national economy, engineering, military, art and social relations amount to problems of that kind.4 At the same time, the up-to-date apparatus of vector optimization is still far from being perfect and continues to develop. Some theoretical results may be employed here as an instrument for argumentation on the decisions made. We isolate the following three main groups of methods for the solution of vector optimization problems15'16: - optimization of hierarchical sequence of quality criteria; - definition of the set of unimprovable points; - definition of the solution based on a particular trade-off. We consider in greater detail the first group of methods, discussing them from the standpoint of methods for handling the criterion priority. The priority of criteria may be specified by different parameters, from various points of view and with different accuracy. In this case the principal characteristics are as follows: a priority series J, a priority vector A = ( A i , . . . , A m ) and a weight coefficient vector (a weight vector) m). The priority series J represents an ordered set of local criterion indices
Q = (ai,... ,a
(44.1) We postulate purely qualitative relations ol dominance (preference relations, including also the binary ones) of criteria, thus the criterion yi is more important than y%, yi is more important than 3/3, etc. and we write them as: (44.2)
The quantitative side of dominance is not included here. In the criterion set, we may find groups of equivalent criteria, denoted by {yj ~ y]+i ~
386
Systems Optimization Methodology
i/j+2}. If such groups are met in the priority series we isolate them with internal brackets, thus we have (44.3)
In specifying the priority series, the following three main cases are possible: - one priority series is specified without equivalent groups (44.1); - one priority series is specified with equivalent groups (44.3); - several variants of the priority series {Jv} are specified. The latter is associated with the presence of several points of view as to the importance of criteria. It should be understood that different interpretations of orderliness generally lead to quite different notions of optimality. The priority vector A = ( A lt A2,..., Am) represents an m-dimensional vector with the components A_, -- binary relations of the priority (Aj e [l,oo]). Here is reflected the priority quantitative characteristic, the binary relation of priority Aj determining the degree of superiority, by importance, of two adjacent criteria from the priority series. If there is t/j > j/j+i, then Aj determines the degree of superiority by importance of the jth criterion over the (j + 1) criterion. Aj has the following meaning: let the criteria
2/j and TJJ+I be normalized, i.e. their scales be reduced to a unified scale (commonly dimensionless), then the solution x is better than x ' f x > x' / if the following condition is met
MM*) -2/7CO) > (l/j+i(*') -2/j+i(*)) • In the case of equivalence of the criteria y} ~ j/j+i, A = 1, we set A m = 1 to carry out calculations conveniently. In specifying the priority vector, the following cases are possible: - A is accurately specified; - A is specified approximately as confidence intervals &_, < Aj < Ojj - A is specified approximately as several qualitative groups: {Aj > 1}; {Aj > a}; {Aj > 6}; {Aj > c}, and the like, here a > 1; 6 > a; c > b are the corresponding gradations of superiority; - A is specified as dependence of Aj on the number in the series of preference
Aj - M);
Methodology of Vector Optimization
387
- for A we prescribe the range A € A; the pairwise comparison of criteria allowing A_, to be determined with adequate accuracy. The weight vector a = (ai,a2,-• • ,am) represents an m-dimensional vector determined in the region a e A = (a\aj E [0,1], £ aj' = 1. j € J}. Each component a,- has the sense of the weight coefficient characterizing relative superiority of the j'th criterion over the others. The priority vector is related to the weight vector by the relations:
The vector a can be determined accurately or approximately, e.g. in the form of specification of a particular region in A or as a hypothesis of the character of changes of Oj from j, the procedure of specifying the weight vector being more complex than that for the priority vector. If we need to specify simultaneously m numbers from A, then in determining A, we determine successively each component without any relation to others, that is why it is advisable to determine A rather than a. The vector optimization problems may have the following forms of specification of the priority characteristics: (1) j7; (2) J, A; (3) J, \3 = if>(j); (4) A; (5) A or a. The procedures of priority incorporation may be quite different, though in transition from one of them to another, the principle of optimality is always corrected. Here are two possible approaches to handling the criterion priority — the principle of rigid priority and that of flexible priority. We consider the methods of handling rigid priority. The principle of rigid priority relies on the criteria being arranged by importance in a series of the priority J; y\ > 3/3 > • • • > J/m, on the basis of which we carry out successive optimization of criteria. The principle of sequential optimization (optimization of hierarchical sequence of quality criteria) based on rigid priority consists in preventing the level of less important criteria to be raised if it brings about even insignificant decrease in the level of a more important criterion from the priority series.8 The idea of such an approach to the solution of the vector optimization problems was advanced in Ref. 23. According to this notion, we first search for a local optimum for the most important criterion, which is fixed as an auxiliary condition of the problem. We next determine the local criterion of the criterion second in importance, though for a new set of feasible programs; the
388
Systems Optimization Methodology
procedure is repeated. In this way a feasible set is successively restricted to a unique optimal solution or optimal subset.* The described principle of ordering vector sets is assumed to be known as lexicographical.8 Lexicographical optimization problems are identified as multicriterion problems with criteria ordered strictly by importance. The lexicographical optimization problems have been carefully studied in Ref. 13. Let U be a set of all strategies of behavior (solutions), the efficiency of which will be characterized by specific criteria ki, k?,..., ks forming in totality a vector criterion. Definition 1: Strategy u is not worse than strategy v in the sense of vector criterion (u > v) if the following inequalities are satisfied: M«)>M").
r = l,2,...,3.
(44.4)
Definition 2: If k(u) = k ( v ) , i.e. s in equalities of the form kr(u) k T ( v ) hold, then the strategies u and v are identified as equivalent to (u ~ v). Equivalent strategies should be regarded as such in the sense of the criterion k. Definition 3: The strategy u is explicitly better and preferable to the strategy v (u > v) if there are inequalities of form (44.4), even one of them being strict. Definition 4: The strategy v is termed efficient if there is no strategy u such that u > v, i.e. for which the inequalities of form (44.4) might be satisfied, if only one of them is strict. In the determinate case, i.e. in the absence of any random and indeterminate factors, the chosen strategy unambiguously determines the result of operation. Therefore, in the determinate lexicographical problem, each strategy u is characterized by s numbers — values of specific criteria: ki(u),k-2(u),...,ka(u). *Here a quasioptimal approach is employed. At each stage we carry out quasioptirnization, i . e . the search for a domain of decisions close to optimal — a quasioptimal set
rather than a unique exact optimum.
Methodology of Vector Optimization
389
All of the specific criteria forming the vector criterion k\ = (k\,..., ka) are strictly ordered by importance, as a result, in comparison to the pair of strategies we primarily employ fci, the first of them, and the strategy, for which the value of this criterion is larger, is assumed to be better. If the value of the first criterion proves to be equal for both strategies, then the second criterion is employed and the strategy, for which its value is larger, is preferred; if the second criterion does not allow the better strategy to be isolated, then the third (specific) criterion is brought into use and this process proceeds up to the criterion ks. If the values of each specific criterion for the strategies under discussion turn out to be equal, then these strategies are referred to as equivalent in the sense of the vector criterion k. Definition 5: The strategy u*, which is not worse than any other strategy lex
v in the sense of relation > (lexicographically not less preferable), is called lex
lexicographically optimal if u* > v is satisfied. Let v' be a set of all lexicographically optimal strategies. Furthermore, let all these strategies be equivalent among themselves so that the optimal value of k* of the vector criterion k is given as k" = ( k i ( u * ) , k 2 ( u * ) , . . . , k s ( u * ) ) , where u' is any strategy from V* The lexicographical optimization problem consists in finding optimal strategies: we can restrict ourselves, almost always, to finding merely the optimal, unique strategy rather than the entire set V* just as in conventional optimization problems since all the above strategies are equivalent in the present case. The definition of optimal strategy implies that the set V* may be specified by the following recurrent relations:
(44.5)
The latter relations imply V D Vf D V2* D • • • D V*, i.e. each subsequent specific criterion restricts a set of strategies obtained through the use of the preceding specific criteria. Thus, for example, if for the original optimization problem with one scalar criterion there are several solutions and for further selection the auxiliary criteria are successively employed, then the strategies resulting from them will be optimal for the corresponding
390
Systems Optimization Methodology
lexicographical problem with the vector criterion composed of all these
alternately employed criteria. In the general case the set V may prove to be empty. It is common knowledge that optimal strategies may not exist also in the ordinary extremal problems with one scalar criterion ki, and its solution then is taken as the maximizing sequence {uk} C V determined by the condition
The notion of the maximizing sequence turns out to be extremely useful also where optimal strategies exist since for the analysis and numerical solution of extremal problems, extensive use is made of iterative methods. The notion of the maximizing sequence also covers the lexicographical optimization problem. In the absence of the optimal strategy, the least upper (lexicographical) bound of the criterion k, for which the designation k* will be preserved, may also be determined by relations (44.5) since for the empty set we have sup kr(u) = —oo. In the general case — oo and +00 may u&v; represent the components of the vector fc* (here +00 can be met but not more than once). We assume that the sets V*, beginning with a particular r = p > 1, are empty. Definition 6: The sequence of strategies {uk } is termed lexicographically maximizing if, for sufficiently large numbers fc, the strategies uk enter into the set Vpl, (so that with p > 1, kT(uk) - k* for r = 1, 2,... ,p-1) and the equality lim kp(uk) — k* is satisfied. k —»oo
Definition 7: If the set of optimal strategies V is non-empty, then the sequence {uk} is identified as lexicographically maximizing and all of its terms, perhaps with the exception of Vsf_1 their finite number, are the elements of the set.
is satisfied for this sequence. Definition 8: The sequence of strategies {uk} is called lexicographically maximizing if for any vector g £ Rs, which is lexicographically less than k*, we obtain the number N ( g ) such that with any k > N ( g ) the vector k(uk] is lexicographically larger than g.
Methodology of Vector Optimization
391
Thus, the lexicographical optimization problem consists in finding the lexicographically optimal strategies or the lexicographically maximizing se quences of strategies. The stated problem is also termed the lexicographical optimization problem and, by employing the notation k* = lex sup k(u), it is sometimes briefly written as follows: find
lexsupfc(u). lex sup k(u).
(44.6)
In the case of existence of lexicographically optimal strategies, problem (44.6) may be written as follows: find
lex max fc(-u). k(-u). u€V
Sometimes the lexicographical problem is more conveniently stated so that all specific criteria should be alternately minimized. In this form, the prob lem represents a lexicographic minimization, and it can be presented as follows: find lexinffc(u) (44.7) lexinfJfc(u). u£V uev It is to be noted that there is the relation: lex sup k'(u) = — lex inf k(u). The lexicographical problem will have a solution and a unique one at that (the set V* consists only of one strategy), if some criterion kT changes to the maximum on V*_1 merely at one point. For example, the problem cannot have more than one solution if V is a convex set, allfcx>$2>---,kT-1 are quasi-concave (the sets V*, V^, ■ ■ ■, Vf—i being convex), and kr is strictly quasi-concave. The dependence of solutions of the lexicographical optimality problem on parameters (including the initial data) is complex and nonregular (nonsmooth), and this should always be kept in mind when considering specific lexicographic problems, especially non-discrete ones. Of practical interest is the possibility of transforming the vector criterion of the lexicographical problem. In this regard, we should refer to some very simple, but quite useful results.
392
Systems
Optimization
Methodology
We assume that the function / has been defined on the set k[(v) of values of the criterion ki and is strictly increasing (i.e. for any x and y from kt(v) such that x > y, there is a strict inequality f(x) > f(y))- Then the vector criteria k = (fcx, &2,..., fcs) and (k\,..., fc;_i, f(ki), ki+\,..., ks) are lexico graphically equivalent, i.e. generate on V the same lexicographical relation lex
of preference > , specifically, the sets of optimal strategies isolated by these criteria coincide. The latter enables one, say to construct, for the problem under consideration, new vector criteria lexicographically equivalent to the original, though more convenient in calculation. We consider the problem of finding lexicographically optimal strategies. Prom the computational point of view, it is rather difficult to solve the lexicographic problem immediately by recurrent relations (44.5) since at each stage, perhaps with the exception of the latter, we need to completely construct the set V*. It is more convenient to solve the following sequence of problems: (1) Find sup fci(u) = k{ u£V
(2) (s)
Find Find
k2(u) = &■> sup fc2(") uev _^I
(44.8) (44.8)
sup ks(u) = = fc* fc* u€V r=l,2,...,s-l
The solutions of problems (s) are only all optimal strategies, but for con struction of constraints in problem (s) we need to know merely s—1 number k*, for which it suffices to successively find one solution at a time for prob lem (1), (s-1). If the (/) problem is the first in which the upper bound is not attained (this is possible if and only if V* is empty, and V*_Y nonempty), then any sequence composed of strategies satisfying all constraints in the (/) problem and maximizing ki, is lexicographically maximizing. In this case the sets of strategies satisfying the constraints in problems ( ( / + 1 ) , . . . , s) are empty, thus k* = - c o for all r = l+l,..., s. The above multistage approach basically allows the lexicographically optimal strategies to be found, though it is rather complex since to solve s problems of optimization is rather bur densome. This procedure has one more appreciable disadvantage: addition of constraints kT(u) — k* to problem (1) in relations (44.8) may result in the loss of its features and, as a consequence, in the impossibility to employ
Methodology of Vector Optimization
393
efficient computational algorithms. As far as the foregoing is concerned, the problem arises with respect to the possibility of a single-stage solution of lexicographical problems and its reduction to one extremal problem with the scalar criterion.13 The study18 shows that the lexicographic ordering of criteria determines one of the possible procedures for convoluting the vector criterion. It is established that a unified criterion can be formed as the weighted sum of specific criteria with a definite relationship of weight coefficients. These results imply the inverse inference: if the unified criterion represents the weighted sum of specific criteria, then, under some conditions, the behavior of the operating side will be just the same as in the case of the lexicographically ordered criteria. The literature provides many examples of analysis of multicriterion problems as well the choice of their adequate procedure for convolution of criteria.1'5'11'13'14 Reference 3 contains the theorems of the unitility theory, from which it follows that many criteria cannot be always reduced to one criterion in principle. Following the work, we will consider the lexicographical problems of optimization under risk. In this case we know the distribution of random parameters, which enables us to determine the probabilities of appearance of possible results of an event or operation. It is possible to distinguish two types of specific criteria in the problems of optimization under risk. As in deterministic problems, the criteria formally are the ordinary functionals determined on the set of strategies, so that each strategy is characterized, through the use of such criterion, by a fairly definite number — value of this criterion. Such criteria, however, have a probabilistic sense: they evaluate the probabilities of emergence (or, on the contrary, nonemergence) of some events or operations, e.g. the probability of no-failure operation of an apparatus or system as a whole. For the fixed strategy, the criteria of another type are the random values with distributions described by the known functions or distribution densities. The specific criterion of that kind characterizes the strategy by a set of numbers rather than by one number. In general, it is inconvenient to compare strategies with the aid of such criteria, therefore they should be transformed firstly by the procedures extensively applied for the scalar problems of optimization under risk.
394
Systems Optimization Methodology
The lexicographic problems under uncertainty amount to finding maximin strategies or a sequence of strategies being lexicographically maximizing for the criterion k ( u ) . The basic principles of optimality suggested for decision making under uncertainty, are carefully analyzed in Refs. 7, 12, 13, 17, and others. We will consider the methods of handling the flexible priority, whose principle presupposes specification of quantitative characteristics of the priority A or Q, which enables us to partly prefer more important criteria when choosing a solution. In practice, this leads to estimation of the quality of
solution by the weighted vector criterion — vector pair (y,a) and to correction of the optimality principle. Basically, this results in transformation of the criterion space as well as in a corresponding change in the scale of criteria. Then comes the choice of the optimal decision based on one of the possible principles of optimality, though in a transformed space of criteria. The principle of flexible priority enables us to give preference, within reasonable limits, to more important criteria with regard to the degree of their relative importance, which should be placed among the merits of the given principle. However, definition of the numerical characteristics of the priority is associated with considerable difficulties. In the approximate specification of the priority characteristics, we have to make recourse to solving parametric problems and suggest in some cases several variants of optimal decisions.8 To obtain the data necessary for the justified decision making, Ref. 6 offers to carry out zoning-partitioning of the set of the nature state vectors into subsets, for each of which a definite alternative solution is optimal. It has been noted that stability of solutions is inherent in the problems with a finite number of operations and an uncountable set of state vectors: with changes in the state of nature or in distribution of the state probabilities, which may lead to changes in the result, a definite decision remains unaltered until a corresponding vector of the nature state moves inside the same subset. Upon further movement, the nature state vector corresponding to actual situation may pass into another subset and cause the necessity of changes in the decision. Stability of the decision shows that the requirement for complete knowledge of the nature state vector is excessive: to make accurate decisions requires merely the data allowing the determination of which of the subsets the actual situation corresponds to. If the initial data are so approximate
Methodology of Vector Optimization
395
that we can indicate merely a group of subsets to which they correspond, then an accurate choice of decision is not ensured, but in many instances it is possible to carry out an approximate selection of decision, where maximal losses do not exceed a specified admissible limit. The principle of rigid priority can also be implemented in the form of an equivalent scheme of the flexible one. To do this, we introduce the priority vector with the components that ensure the choice of decision based on the rigid priority of criteria. 8 References 2, 4, 13 offer the method of successive concessions for the case where all objectives are ordered by their importance. Following the study, 13 we assume that the criterion ki is important for definiteness, k2 is less important, and then come other specific criteria feg,&4,... ,fes according to the degree of their importance. Let the criterion k\ and the first in importance at that be maximized, its largest value Q\ being determined. We next assign the value of the "admissible" decrease (concession) Ai > 0 of the criterion k\ and the largest value Q2 of the second criterion k2 provided that the value of the first criterion must not be less than Q± - Ai. We again assign the value of the concession A2, by the second criterion, which, together with the first, is utilized in finding the relative maximum of the third criterion, and so on. Finally, the criterion ks, which is the last in importance, is maximized provided that the value of each criterion from the preceding s — 1 should be not less than the corresponding magnitude QT — A T ; the resultant strategies (solutions) are assumed to be optimal. Thus, we take as optimal any strategy being the solution of the last problem from the next sequence of problems (assuming that kr are bounded from above). The latter ensures finiteness of all magnitudes QT, and consistency of constraints in all S problems with any A r > 0. We have: (1) Find Qi = sup k\{u) uev uPV (2) Find Q2 = sup k2(u) uev Mt*)>Qi-Ai (44.9) M«)>Oi-Ai (s)
Find
sup ks(u) = Qs uev kT(u) > Qr — — Ar, Ar,
rr = = 1, 1, 22,,......,, ss — — 11 ..
396
Systems Optimization Methodology
If the criterion ks on the set of strategies obeying the constraints of problem (s) does not achieve its largest value Qs, then the solution of the multicriterion problem is thought to be the maximizing sequence of strategies {uk} from the above set ( lim ks(uk) = Qs). fc—i-oo
It is advisable to consider similar maximizing sequences also for the case wherein the upper bound in problem (s) is achieved since extensive use is made of iterative methods for the solution of extremal problems. Comparison of problems (44.9) with problems (44.8) shows that in the case of all A r being zeros, the method of successive concessions isolates only lexicographical optimal strategies, these strategies delivering the value largest on V to the criterion ki which is the first in importance. In the other extreme case, where the values of concessions prove to be very large (e.g. such that Qr — A r < inf ki(u), r = 1,2,... ,s — I), the strategies obtained uSV
with the aid of this method, delivering the value largest on V to the specific criterion ks, which is the last in importance. The values of concessions assigned for multicriterion problems, may be regarded as a peculiar measure of deviation of the priority (a degree of relative importance) of specific criteria from the rigid, lexicographical priority. The approach to the process optimization problem depends basically on the possibility of estimating the measure of unequivalence of different specific criteria and specification methods for this unequivalence. The process optimization problem may be most easily solved if all specific criteria can be arranged by importance. We now develop the idea of the hierarchical scalar criterion ranking in the optimization problem. Let the mathematical model of the controlled system be specified, i.e. we specify: (1) the equation of movement in the vector form
x = F(x,u,t); x£Rn,u&Rm; (44.10) (2) the domain of its distribution
N(x,u,t)>0; (44.11) (3) the time interval r = (0,T) or T = [0,oo]; (4) the set V of piecewise continuous functions assuming the values from (44.11), which is identified further as the set of admissible controls;
Methodology of Vector Optimization
397
(5) the function F(x,u,t) smoothened by (x,u) in the domain (44.11). Suppose we have specified the boundary conditions for the problem, and we denote them symbolically by the equality (»,/) = 0.
(44.12)
Furthermore, let the vector function be specified as
J(u) =
= <{>a(x,u,t), (a = l,...,fc). (44.14)
On any feasible trajectory x(t), u(t) each function Ja(u) assumes a definite numerical value and, consequently, J(u) represents a vector functional. We will consider the first, scalar component (f>i(xtu,t) of the vector criterion for optimization (44.13). Let the control w' 1 ' transfer the minimum to this
functional and assign the value <j>\ = c(>i(x,u^,t) to it. With u'1', we calculate the values of the remaining criteria 4>^ i4>z\ • • • ,4>k Let tne control u^ transfer the minimum <j>T to the criterion >2, thus
cj)k '. Iterate this operation till all scalar criteria are exhausted. After completing the operation, we obtain the matrix
(44.15)
which will be identified as characteristic. As the matrix is finite, the analysis of numerical values of its elements and estimation of the gain (or loss) in values of different system quality indices enable one to make a definitive choice of the basic criterion. In this case the remaining criteria will obey hierarchically. The hierarchical ranking idea suggested here, is alluring because of its simplicity for application in the simplest cases to solutions of the optimal trajectory programming problems. It can be useful wherever the characteristic matrix points to weak sensitivity of criteria to controls and reveals an
398
Systems Optimization Methodology
explicit advantage of a particular criterion over the others. Such an advantage enables one to draw a conclusion about hierarchical subordination of criteria. The difficulty in choosing a preferable criterion is preserved in all other cases where the optimizing criteria are weakly sensitive to the choice of a control, and the matrix elements entering into the same column will not undergo substantial changes associated with those in their upper index, i.e. in control function (44.13). Investigations in this direction have been carried out in Refs. 21, 23, 25, etc. In Ref. 25 the basic criterion is obtained from the engineering analysis, and other criteria are stated as hierarchically subordinated to this criterion. Reference 13 studies some approaches to the solution of the process optimization problems by several successively utilized criteria as well as the available methods of process optimization by scalar criteria. Specifically, the problem of utilization of the Pontryagin maximum principle has also been discussed. In so doing, the lexicographical optimization problems for continuous and discrete processes were first studied, and then the method of successive concessions was analyzed. Optimization of solutions in hierarchical systems or in their relation (the synthesis problem) gives rise to considerable difficulties, which is determined by special features of hierarchical systems: distribution of subsystems by different hierarchical levels, subordination of one level to another, multiplicity of objectives and other conditions for functioning. As a consequence of diversity of the nature of the vector optimization problems, it seems possible to largely overcome the above difficulties by utilizing formalism of the vector optimization.8 The above paper discusses the method described below. The behavior of the hierarchical system features a series of parameters. As first we consider the behavior of the system without conditions determining the hierarchy. We then consider the implemented problem with conditions determining the hierarchy. This enables us to carry out natural decomposition of the general problem into the vector optimization subproblems of two classes — the problems of optimization at each hierarchical level and the interlevel optimization problem. Next, as a result of convolution of vector criteria, each hierarchical level obtains a generalized scalar criterion, which enables one to solve the second class of subproblems — interlevel optimization. The latter problem applies to the class of vector optimization problems with priority. At first it must be solved conceptually. Convolution of the vector criterion is carried out with regard to the priority vector.
Methodology of Vector Optimization
399
With considerable differentiation of priority relations, algorithmic solution of the lexicographical or quasilexicographical ordering type is also possible. Thus, the formalism of vector optimization enables one to construct a well-founded procedure of selecting the optimal solution in hierarchical systems. In this case the complex, original optimization problem reduces to the scalar optimization problem or a sequence of such problems.
§45. FINDING THE SET OF UNIMPROVABLE POINTS In implementing the vector optimization problems, it is common to firstly isolate the tradeoff domain (the set of unimprovable points, Paretooptimal solutions) since in the majority of vector optimization problems there is a certain range of solutions, where the criteria are consistent. The tradeoff domain obtained is presented in a number of cases as an optimal solution to the problem, e.g. "negotiation set" in the theory of cooperative games,3 "efficient programs" in the Koopmans economic theory,10 and the like. We will consider the definition of the unimprovable system which has been first given by L. Zadeh.16 Let C denote a subclass in E determined by a set of values of the system S. Introduce the following notation:
S(S) - a set of all systems, each of which is better than the system S. E(5) - a set of all systems, each of which is worse than or equivalent to the system 5. E(5) - a set of all systems, which cannot be compared with the system . . . Each system from E belongs to one of these sets given above, and the linkage of these sets determines S. Definition 1: The system 50 6 C is unimprovable in C if an intersection of the sets C and E(5o) is empty. In other words, this means that in C there is no other system which might be better than system 50. The present definition is also correct for the case where E is determined by the vector index of quality. Thus, for example, let S be determined by the vector values x{xi,... ,xn}, and C € Rn• Next it is assumed that the quality indices of the system S are measured by components of the vector 1/3(1) {y>i(x),... ,f>m(x)} wherein each (Pi(x) is the real-valued function from x.
400
Systems Optimization Methodology
The system 50 determined by the vector value z0 € Rn, is unimprovable in Rn if for any x € Rn there is ^(IQ) < Vt(^)> (* = l , - - - , T n ) (the minimization case of all quality indices
* = !,...,tn,
where x(x\,... ,xn) is the system state vector, U(HI, ... ,um) the control vector. The functions fi(x,u,t),... ,fn(x,u,t) are continuous by z, u, and continuously differentiable by D. We specify x(ti), and the time interval T = {t : t\ < t < t2}, which completely determines x(t) with the specified u ( t ) . It is required that the control function u(t) should satisfy the following three conditions: (a) u(t) belongs to the class of operative convex functions C at the interval T; (b) the trajectory x(t) corresponding to the control u(t) must stay in the
feasible domain X, i.e. x(t) € A" with any t € T; (c) the trajectory x(t) must take on the given value x(t%), where t% may be fixed or arbitrary, satisfying t^
The acceptable control A is thought to be worse than the acceptable control B if there is the relation
y*(ti}\A >t(ti)\B y in which the sign of inequality is preserved at least for one value of i. The acceptable control is identified as unimprovable if it is not worse than the remaining acceptable controls. The unimprovable acceptable control determines the unimprovable system. Theorem 1: Specify the fixed points x(t\) and x f a ) , and let X be unbounded by the X-space. For the acceptable control u(t) and its corre-
sponding trajectory x(t) to be unimprovable, it requires the existence of
Methodology
of Vector
401 401
Optimization
the vector of the function $(^>i,..., tpn) and the function H(ijj,x, u, t) sat isfying the following expressions: n
H(4>,x,u,t)
N
=
Y2^{t)h{i,u,t)~Y^ckkgkk{x,u,t), (x,u,t), i-l
k=l rt2
45 1 (45.1) C -)
du /I' H{i,,x,t)dt
dipAt)
dH
,
,
,
,
In the latter expression (45.1), the inequality sign cannot be preserved for all k values; we denote in terms of Su the first variation of the next integral corresponding to the infinitesimal variation u(t) with the fixed f», x, t. The theorem stated above is titled "the general theory of optimal control". Its proof has been constructed in Ref. 6, which also gives the mathematical results stated as necessary conditions for unimprovable controls. However, the above general theorem and the subsequent ones from Ref. 6 cannot be successfully applied to solve the particular control problems with many criteria of quality since the problem of determining the coefficients ck, k = 1,7V in (45.1) remains to be solved.4 The general vector optimization problem can be stated as follows.2 Let X be the solution defined on the feasible solution set X. The quality of solutions is estimated by the local criteria y\, 3/2, •■ • forming the efficiency vector y = (1/1,2/2, ••• ,Vm)- The vector y is associated with the solution mapping x —► y = f(x) specified analytically, statistically and heuristically. The relative importance of local criteria is given by the priority vector A = (Aj, A 2 , . . . , A m ), where Xj G [l,oo) is the binary relation of the j t h criterion priority over the (j + 1) criterion. We need to find the optimal solution x° determined by two conditions: - the solution must be feasible, i.e. belong to the feasible solution set X; - the solution must be the best, i.e. optimize the efficiency vector y with regard to the criterion priority A. The present formulation corresponds to the general vector optimization model. M o d e l 1. 11 v°=fx° =f~\opt(y(x),\) \ opt (y(x), A)l
(45.2)
402
Systems Optimization
Methodology
or in the non-uniqueness case of the optimal solution, wherein the optimal solution subset X° is isolated as optimal, there is also the model of a more general type: (45.3) where opt is the optimization operator determining the principle of optimal-
ity and having the sense of the order relation; f~l is the inverse mapping
y ->z = f~l(y}. Model 1 is the vector optimization model since it has the criterion of quality (or efficiency) being the vector, and its implementation results in finding the optimal solution i° or the optimal subset X°. In vector optimization problems there is no contradiction between local criteria. This contradiction is commonly non-strict. There is a certain
domain of agreement Xs which has no contradiction, and the solution quality may be improved simultaneously by all local criteria. At the same time, we have the tradeoff domain Xc where there is contradiction even with one of the criteria. Here improvement of the solution quality by some criteria brings about deterioration of others, and the choice of any solution is based on tradeoff. Thus, the range of feasible solutions consists of two nonintersecting parts
X=XCUXS, XCC\XS = 0. (45.4) The optimal solution should not belong to the agreement domain Xs since any solution from this domain may be improved by all criteria. Therefore it will be, without fail, in the tradeoff domain x° € X, and the field of search for an optimal solution should be confined only to this domain. The problem thus arises with respect to determining the tradeoff domain, and its isolation from the range of possible solutions. In models (45.2)-(45.4), this involves restriction of the range of feasible solutions and transition to the vector optimization problems with a strict criterion contradiction and new ranges of feasible solutions Xc and Yc. Definition 2: The agreement domain Xs is identified as a subset of the
set of feasible solutions X featuring the possibility for each solution x £ Xs to be improved without lowering the level of any one of the local criteria.
In other words, if x £ Xs, then we always find in X a better solution x',
Methodology of Vector Optimization
403
for which y(x') > y ( x ) . The relation > means that for all j, yj(x') > y(x) and even for one j, y(x') > y ( x ) . Definition 3: The tradeoff domain Xc is identified as a subset of the set of feasible solutions X featuring that solution x € X° cannot be improved without lowering the level even of one of the local criteria. Thus, if x, x' 6 .,YC, then in the case of two criteria y i ( x ) > y i ( x ' ) , y^(x) < y i ( x ' ) or vice versa. In other words, for any two solutions from Xc we have without fail a contradiction of criteria, and the choice can be made only on the basis of tradeoff. The points x1 E X are also referred to as effective points, unimprovable solutions or Pareto-optimal solutions. The trade-off domain can be defined in two basically different ways:
- by defining the agreement domain Xs and eliminating it from the range of feasible solutions
Xc = X\XS (45.5)
- by isolating the trade-off domain proceeding from its properties (Definitions 3, 4)
(x,y)^Xc (45.6)
The second way is preferable. It is associated only with one operation. Furthermore, it is common that Xs > X° and isolation of elements is more complex than X°. The model of choosing the trade-off domain, which is common to all classes of vector optimization problems, corresponds to Definition 3. Model 2:
Xc = {x\x e X, {x'\y(x') > y(x)} n X = 0} (45.7) or in the space of criteria Yc = {y\y € r, {y1 > y] n Y = 0} .
(45.8)
At the basis of this model is the principle of optimality opt y = max?/, i.e. that of dominance: the solution x is better than the solution x' if y(x) > y ( x ' ) .
404
Systems Optimization Methodology
For purposes of calculation, it is more convenient to employ the parametric model of optimality equivalent to (45.7). Model 3: (45.9)
Here a = ( a i , a 2 , . . . ,a m ) is the vector parameter determined on the set 771
A = {a\a > 0, ^2, aj = !}• For convex problems (where A is convex), >=i implementation of Model 3 requires to find the local maximum of the linear
form L(x) = Z^jl/jW in modifying a € A. For non-convex problems, we i need to find all local maxima of the linear form L(x) and test them for dominance conditions. We consider some properties of the set of unimprovable points as well as the relation between the optimization procedure for hierarchical sequence of quality criteria and the procedures of defining unimprovable points following Ref. 1. We assume that the system behavior is characterized by the rc-dimensional vector x — {xi,... ,xn] from a certain given set X of the Euclidean n-dimensional space En, and it is estimated by the m-dimensional vector function
i = l,m; £Z Aj — 1 having the sense of relative utility of the zth i=l
component of the criterion (general utility is taken as equal to a unity), and we choose, as the problem solution, the vector x° £ X minimizing m
on X the function f\(x) = J^ AJ(/J;(:E). 1=1 (2) Optimization of the hierarchical system of functionals.
Among the
vector components ip(x), we specify the order of preference. To be specific, we take the following order:
V^ (x), 3, 2 ( x ) , . . . , <^m ( x ) .
(45.10)
Methodology of Vector Optimization
405
As a solution, we assume the vector x° £ X minimizing the hierarchical system of functions (45.10), i.e. x° satisfying the relations: 00 (fii^x fi.ix )) = = minL ^ t , ( x ) ,
x£X°
tpl2{x°)
min. m i n
min1
1
lex" - 1 16X"'-
where X° = X, Xs = ix ; ^ , (s) =
min ^
(*), x e .Y- 5 " 1 1.
(3) Definition of unimprovable points. We introduce the notion of the set of points unimprovable in X with respect to ^p(x). The choice of unimprovable points is thought to be judicious with the vector criterion of quality. The set of unimprovable points was studied in Refs. 7, 9, 16, and others. The work in Ref. 11 discussing some properties of the set of unimprovable points in the general dynamic problem with k criteria of quality has shown that the problem amounts to minimization of the linear form of the vector components
ip\{x) = y]Xjifij(%), V ] Xnpi(x), tpxjx) t=i
X{ > 0,
i = l,m;
J] Xj — 1. i
Concurrent with unimprovability, L. Zadeh has introduced the notion of optimization with a vector criterion, i.e. the point x° € X is called optimal in X with respect to
406
Systems
Optimization
Methodology
Problem 1: Find x° e X such that (p\(x°) <
i = l,m, ]T A, = 1}. 2=1
Problem 2: Let H(J) be a permutation specified on the set J = { 1 , 2 , . . . , m}. Find x° G X being the solution of the minimization problem with the hierarchical system of functions, where the order of preference is specified according to 11(^7). Problem 3: Let J\,..., Jk, k < m be a specified, ordered partition of the index set J Specify the numbers A, > 0, i = l , m such that 5Z At = 1, s = 1, k. Find the point x° G X such that (pi(x°) = min 0i(x), xex° 002(z°) ,
A . ^ z ) , 5 = 1, fe, Kvi(x), k, A 0 = A",
s 1 X Js},} 1 A* = {x {x : i e6 Pr - , f^t ,( (j l) )==f j^((i l, !°)), ,»i = X
s5 = I7fc. l,k.
Problem 3 is more general than Problems 1, 2. Indeed, if partition of the index set consists of one set J\ = J, then we get the problem J. If each set Js entering into the partition J is the single-element one, then Problem 3 becomes a variant of Problem 2. The following statement holds true. S t a t e m e n t 1: If x° G X is the solution of Problem 3 for a certain partition J i , . . . ,Jk and some Aj > 0, i G J, Yl A, = 1, s = 1, k then x° is also the unimprovable point in X with respect to tp(x). From this statement we have obtained the following corollaries. Corollary 1: If x° G X is the solution of Problem 1 with a given A, A G A + then x° is also unimprovable in X with respect to tp{x).
Methodology of Vector Optimization
40'i
Corollary 2: If .r° 6 X is the solution of Problem 2 with a specified ordei of preference, then x° is also unimprovable in X with respect to
if the value assumed by
of preference if and only if f(x) = const for all x & N. Statement 3: In order for the point x° optimal in X with respect to
to
Ji,...,Jk of the set J and such A, > 0, i e J, XI A, = 1, s = ITfc with
iej, which x° is the solution of Problem 3. By employing the theorem of detachability of convex sets, we will study the sets of unimprovable points. To do this, we need to utilize the following
408
Systems Optimization
Methodology
auxiliary statements. Let fi», i = 0 , 1 , . . . , m be the bounded, closed convex sets from En m
L e m m a Is In order that f| ft, ^ 0 , it is necessary and sufficient that i=0
(45.12)
x j ' ( a ; ) > < 00.. max < min V > o a,, ) xi - -V m > a maxo'(x) ||s||En,„=i | * e n 0 \{r{ ) ^*en< J Here g g £ n m , g = {§i,...sgm}, (n x m.) space.
where gl e En, i = U n , £ n m is the m
L e m m a 2: In order that fto C H ft,, it is necessary and sufficient that i=i
max
(45.13) (45.13)
} qQi —>> maxo'(i) maxo'(i) >><< 0.0. < max I > l\x I £—
As before, it is assumed that the functions <pr(x), i = l , m and the set X are convex and, besides, the functions
A(<5) =
f
/m V
1\
o'x >> max < min >^ o, a, \x — V^ max q'.x ||9||E„,„=i | x e x \ ^ ^ y ^ £-l€Gi(fii) j The function A(<5) is continuous and nondecreasing in 6. We now consider the equation (45.14) \{8) \{6) = = Q. 0. (45.14) Introduce the concept of unimprovable roots (45.14). We say that 8° is the unimprovable root if A(<5°) = 0 and there is no 8 such that A(<5) = 0 and <5j < 8°, i = l,m, even one inequality being strict. Among unimprovable roots of this equation and unimprovable points there is a close relation revealed by the following statements. S t a t e m e n t 6: In order for the point :r° to be unimprovable in X with respect to
Methodology of Vector Optimization
409
Therefore, the set of values assumed by the vector function ip(x), when x runs through N, belongs to the set of unimprovable roots. Statement 7: If 6° is the unimprovable root, then there is such a point x° unimprovable in X with respect to (p(x) that
where g° = {?,. ..,9^1 is tne vector transferring the maximum in the left-hand side (45.14) with <5 = 6° Therefore, the set of values assumed by ip(x) on N, coincides with that of unimprovable roots. Let us consider the optimal control problem of the linear system described by the system of differential equations with the vector criterion of quality. Specify the following system: (x) = A(t}x + b(t, u),
X(t0)
= x0 ,
(45.16)
where x is the n-dimensional vector, and u(t) the r-dimensional measurable vector function of control adhering to restrictions:
u(t) £ V for alHe [t0,T]. (45.17) V is the compact set from Er, and the function in (45.17) is the ndimensional continuous vector function. The quality of the process is estimated with the vector functional
V(x(T)) , (45.18)
J(u) =
where
feasible control u°(t) is unimprovable for system (45.16) with restrictions (45.17) by vector criterion (45.18) provided there is no other feasible control u(t) such that Ji(u) < Ji(u°), i = l,m, with even one inequality being strict. Let M = {6 : 6 6 Em,6 = J(u),u e N}. The following statement provides constructive description of the set M.
410
Systems
Optimization
Methodology
S t a t e m e n t 8: The set M coincides with the set of unimprovable roots of the equation max
( V g&{ ) 4>(T, t0)x0 + [ m
min ( V g{) <£(T, t)b(t, u)dt
\
— >V^ max max g[x g'-x>>== 0( . (45.19) Here <j>(t,r) is the matrix satisfying the conditions d$(i, r)/dt = A(t)
By employing the auxiliary problem p(x):
f{x), x, x £ X the authors show that x° is efficient if and only if x® is the optimal solution p(x). This result enables one, by utilizing, say the Kuhn-Tucker conditions, easily to find whether the given point is efficient. On the contrary, if x* is the optimal solution p(x), then x" is efficient and fix*) < /(*)• The problem of interrelation between efficient points proper is also discussed. To find efficient points in the multicriterion control problem 12 suggests the method of internal penalty functions. The problem is solved for objec tive functions: /*<«) = {x(b) - &, W%[x(b) - ft]) + + I (£&) -
a(t}x(t)Qi(t)
J a
x [[xXit(t) -- a(t)x(t)])dt a(t)x(t)])dt++ [[ 6t(t)H*)lI**, 6t(t)H*)lI**, (» (» == TO TO with constraints x(a)=x0,
x = Ax + Bu a
Methodology of Vector Optimization
411
Here the scalar function bj(t) > 0, vectors Xi(t) and & and matrices Wi, c
r(t)i Qi(t) are specified; u(t) is the vector function measurable by Lehbeg on [a,6]. The efficient controls u* are such that, for any feasible controls it the system of inequalities /,(u) < fi(u*), (i = l , n ) , from which at least one inequality is strict, is inconsistent. In the absence of constraints on u the problem is solved by common minimization of the scalar criterion n
n
fa(u) = 51 aJi(u)> where al > 0 and £} at = 1. The explicit form ua i=l
i=l
u* is extracted. Under restrictions of form \\u(t)\\ < I with the aid of the penalty function the problem is reduced to minimization f " ( u ) = fa(u) + b
f \\u(t)\\2kdt. We extract the explicit form of controls w£ minimizing /£ and state that £«) -» f , ( u a ) , k -> oo. References 5, 8, 13, 14 and other studies deal with finding unimprovable points for multicriterion linear programming problems. Reference 8 discusses the problem of finding the optimum of the Pareto objective function ex (where x € Rn, c is the k x n matrix) on the polyhedron F determined by the system Ax < 6, x > 0. Let the point x° 6 F bre specified. Denote in terms of Pxo the problem of maximization 1TS under conditions ex = Js + cx°, Ax < 6, x > 0, s > 0. Theorem 1: If the problem PXo has no finite maximum, then the set of efficient points is empty. Reference 5 systematically studies the properties of the set of efficient solutions in the multicriterion linear problem. Consider the following cases: (1) The feasible set is arbitrary. (2) The feasible set is convex. (3) The feasible set is a convex polyhedron. In conclusion it may be said that a search for the set of unimprovable points is not always needed and well founded in solving the problems of vector optimization. It is not needed in the problems with strict contradic-
tion of criteria. In these problems there is no agreement domain Xs = 0 and the range of feasible solutions constitutes the tradeoff domain (the set of unimprovable points) X = Xc. Therefore, before defining the tradeoff domain, we have to make sure whether the above case exists here. Furthermore, definition of Xc is not necessary if the problem presupposes a search
412
Systems Optimization Methodology
for the optimal tradeoff solution, especially if the algorithm is low-sensitive to the size of the range of search. Exact definition of the tradeoff domain by models 2 and 3 as well as by the Hermeyer models encounters serious computational difficulties. Hence, in a number of cases we may utilize approximation methods. Approximate definition of the tradeoff domain can be accomplished by following the procedures: scanning, directed search, probing, approximation, and the like. All of them are related to common properties of the tradeoff domain and directed towards ordering the choice of the parameter a in model 3, or towards simplification of the procedure of testing the dominance conditions. In a number of cases, where the number of solution variants and local criteria is countable, the vector problem is formulated in the matrix form — as "matrix efficiency," the procedure of defining the tradeoff domain becomes more simplified. Here matrix algorithms are utilized to estimate dominance, which results in coming from the initial set of solutions to the tradeoff set.2
§46. DETERMINATION OF THE SOLUTION BASED ON A PARTICULAR TRADEOFF It is commonly insufficient to be confined to isolating the tradeoff domain to solve specific vector optimization problems. It is generally required to determine a unique optimal solution or at least several solutions. This involves intrusion into the tradeoff domain as well as selection of the optimal solution based on a particular tradeoff scheme and its corresponding optimality principle. The number of various schemes, even in the case of two local criteria, is very large, which aggravates selection of the requisite tradeoff scheme corresponding to the real situation of decision making. Hence, in every specific case, we have to choose the tradeoff domain — optimality principle without fail. Following the study, 7 we will discuss the basic tradeoff principles and their corresponding schemes. For simplicity, we assume that there is a normalized problem without priority. I. Uniformity Principles If criteria are normalized and the same in importance, it is natural to seek a uniform and harmonic improvement of the quality of all local criteria.
413
Methodology of Vector Optimization
In this case the uniformity principle can be implemented in a variety of ways. 1. Equality principle c
oopt p t 22 yy = = {{yi = y y22 = = ■ ■ ■■ ■■ ==ymy}m} £6 YYc.. Vl =
(46.1) (46.1)
Maximization is carried out here provided the level of all criteria is equal. This principle is unduly "rigid." It may lead to a solution outside the tradeoff domain and may even have no solution at all, especially in discrete problems. "In the pure form," the principle of equality is primarily applied in optimization of solutions in cooperative and coalitional systems. 2. Principle of uniformity (maximin) (46.2)
y3 . opt 3 c y = max min p, <=Y yygyc
eY° yy£Y°
Jj
Uniformity of raising the level of all criteria is realized by "pulling up" the worst of the criteria, i.e. those with the lowest level. Apart from uniformity, this principle also has another important meaning — the guaranteed level — min y3. 3. Principle of the best uniformity (sequential maximin) (46.3)
opt4 y = = max mini max min y, y*...... yev j yevc j
y € yc
The notion of uniformity is somewhat augmented as against (46.2). In the non-uniqueness case of solution by the maximin criterion, we define the second minimum and perform its maximization, etc. Thus the principle of sequential maximin is implemented and better matching and fuller ordering of Yc is achieved, i.e. the unique optimal solution is defined. The optimality principles opt 3 and opt 4 can be implemented in the equivalent integral form. m
(46.4)
opt 5 s minc V V . y€Y
—« J= l
4. Principle of quasi-equality c c opt opt66 = = {y\\ { T /yjI I ^- - ^y.\\ H <S,j,ueJ}C\Y <S,j,ue J}r\Y
.
(46.5)
414
Systems
Optimization
Methodology
All local criteria are maximized providing their level will not differ from one another, by not more than the magnitude 6. 5. Quality matching principle Two theorems on average values of the highest steps underlie this principle. S of average degree from the set {yj}, j € { 1 , 2 , . . . , m } is identified as the expression
< 46-6)
M%}=ff>j/mj M%> = ff>J / A
(46.6)
The following asymptotic transitions hold true for them lim As{yj} As{yj}
S—>OC S^OO
= max{j/ m a x ^ }; };; J J
lim As{yj}
S—► — DO S—* — DO
= min{^} mia{|&} .■ J J
(46.7)
For the finite m and y]y equalities (46.7) are satisfied with the finite s > s" (46.4). This property underlies a series of equivalent transformations of vector models and, specifically, construction of equivalent model (46.4). The quality criteria matching model may also be constructed proceeding from this feature m m
s s opt 7 = min y>^ Y~Y~ opt? ,,
(46.8)
where s £ (!,#*}, s* = logm/log(l + e). As the parameter s increases from s = 1 the matching of criteria levels is performed, and with s > s' we obtain the ideal matching equivalent to the sequential maximin model (46.3). However, by choosing s somewhere in the interval (1, s*), we get only partial matching and, therefore, take into account the uniformity principle rather approximately, which offers a means of achieving a high level of general integral efficiency. II. Valid Concession Principles All uniformity principles (46.1)-(46.8) pertain to the category of the "valid" ones. In these principles, the "validity'' manifests itself in the ten dency to match the level of all local criteria. We consider the approach based on estimating and comparing a decrease and increase in the level of local criteria, which are bound to occur in the tradeoff domain. This
Methodology of Vector Optimization
415
principle is identified as the concession estimation principle. It has two modifications: the principle of absolute and that of relative concession. 1. Absolute concession principle c opt V AAy opt88 y= y=\y\y Y] J2 AAg VJ3 >> E ^ 3 \\ nnYCY' , yeY y€Y* II JS+J j€-J * ?€+J J) je-J
(46.9) (46 -9^
where + J7 is a subset of the majorized local criteria (At/j > 0); ~J — a subset of the minorized local criteria (AJ/J < 0); Ayj — the absolute magnitude of an increase. The model is based on the following tradeoff principle: the tradeoff, where the total absolute level of lowering one or several criteria does not exceed that of raising other criteria, is assumed to be valid. This princi ple corresponds to the criteria sum optimization model (that of integral efficiency). opt optg9 y = max ^> y:3 . (46.10) j=J
A serious disadvantage of the absolute concession principle is that it may admit pronounced differentiation in the levels of some criteria, i.e. a high value of the generalized integral criterion can be achieved at the cost of one criterion or a group of criteria with a rather low level of other criteria. However, it is suitable for implementation. 2. Relative concession principle
optio y=\y opt™ y=ly
X! K J - 12 KKJ \ nn yy cc ' ^2 KJ > Yl J \ -
(46.11) 46 n
( - )
where Kj is the modulus of relative changes — "the concession price." Sup pose that all of the local criteria forming the efficiency vector have the same significance. Then we consider as valid the tradeoff where the total relative level of quality deterioration by one or several criteria does not exceed (less than or equal to) the total relative level of quality improvement by the re maining criteria. The model corresponding to the given validity principle, will be identified further as the valid relative concession model. The absolute concession principle is indifferent to the real value of cri teria and, hence, can admit greater uniformity of the criteria levels. There fore it can be utilized only together with one of the uniformity principles.
416
Systems
Optimization
Methodology
The relative concession principle is rather sensitive to the value of crite ria, the value of concession for large value criteria automatically decreasing due to concession relativity, and vice versa. This results in the considerable smoothing of criteria levels. An important advantage of the relative conces sion principle also is that it is invariant to the scale of criteria measurement. If the importance of criteria is not the same, the idea of the valid trade off based on estimation of concessions loses its lucidity, and argumentation of the choice of the weighting vector of importance distribution proves to be rather difficult, especially, where the number of the criteria involved is great. III. Other Principles of Optimality 1. Principle of isolating the basic criterion opt 12 y = max1
(46.12)
where Y1 = {y\y: > yf, j £ {1,2,... ,mj} n Yc. Here we isolate one local criterion as the basic and carry out its scalar maximization providing the level of other criteria is not less than the admissible one. Any criterion may be chosen as the basic, but it is more convenient to take the criterion for which an admissible level is difficult to define. By (46.12), we can imple ment, in principle, any tradeoff scheme and obtain any optimal solution in the tradeoff domain. However, in the overwhelming majority of cases it is impossible to give reasons for the choice of an admissible level of criteria f|. Moreover, it is also impossible to advance arguments for the choice of a particular notion of tradeoff or optimality by varying the magnitudes yf. In some cases, when implementing the given principle and advancing argu ments for the choice of y9, it is wise to utilize equivalent reciprocal models, wherein the basic and additional criteria interchange their positions. 2. Principle of maximizing the weighted sum of criteria opt 13 y = max V V* actjPj }yj ,
(46.13)
where a} € [0,1], Y.jaj - 1 , ] £ J = { l , 2 , . . . , m } . The principle rep resents a modified absolute concession model for the priority case, where the weight coefficients a3 are introduced into the model to offer a relative
Methodology of Vector Optimization
417
advantage for more important criteria in the sense of concession. However this principle has somewhat universal meaning, too. This pertains to the fact that by employing the model, we define the tradeoff domain for con vex problems via variation of weight coefficients. Thus, through the use of model (46.13) it is possible to obtain any tradeoff solution and, there fore, any tradeoff principle is implemented. In practice, it is impossible to advance arguments for the choice of weight coefficients to implement a certain tradeoff principle, therefore the present model plays, in most cases, a specific, local part of the tradeoff model based on estimation of absolute concession in the priority case. 3 . Probability maximization principle of achieving an ideal quality opti 4 y = max m a x pp,, (y > yuy).u). (46.14) yeY" c vev In the stochastic vector problem, we can often define the desired ideal value of all local criteria jtf and, therefore, that of the ideal quality yu In this case, (46.14) with a criterion is the probability of achieving a compound event p(y > yu). Practical methods of calculating the probability of events, even in the case of two or three events, are rather complex. Therefore the present method can be utilized only in some specific cases where m < 3 and calculation of p(y > yu) is carried out simply enough. 4. Optimality principle for multivector problems Multivector problems in the class of vector problems are the most com plex. In these problems, there are several reasons for transition to vector formalism, and, as a result, recourse is made to multivector criteria with components, which are vectors, too. opt y
y€Y v€Y
(46.15)
i.e. y 6 Em x El Such problems are solved by employing a decomposition principle. The general multivector problem is broken down into a series of single vector problems (according to the vector order). Next we solve each subproblem in the ordinary, but only conceptual way, i.e. we define for each j/th subproblem the principle of optimality o p t ' " V The general optimal ity principle results from aggregating the subproblem optimality principles.
418
Systems Optimization Methodology
For each subproblem, the situation has specific features and, in the general case, there is its own principle of optimality. Modern literature outlines two basic approaches to the problem of seeking tradeoff solutions, which are the analytical and analytical-heuristic approaches. Both approaches are based on the central concept of tradeoff axiomatics. By the tradeoff axiomatics we mean a set of axioms strictly determining the tradeoff scheme principle, i.e. the principle on the basis of which the space of criteria is ordered and the optimal solution y° is chosen. In the analytical approach, axiomatics is chosen and the procedure of seeking the optimal solution is implemented following the scheme 5 -> A ->• opt -. y° -* x° ,
where 5 is a certain decision making situation, A the tradeoff axiomatics. Thus, the given situation 5 determines the choice of the corresponding tradeoff axiomatics A on the basis of which we choose the tradeoff scheme opt y and then carry out the search for the optimal solution y° and x°, i.e. the whole process of decision-making is carried out in a strictly formalized way. In the analytical-heuristic approach, transition from 5 to A is carried out manually, analyzing the situation and choosing the most suitable from the suggested set of axiomatics. We will consider both approaches in greater detail. To do this, we define, following Ref. 5, the multiobjective optimization problem as the set: ( T V , x { > } , e N ) , where N = { 1 , 2 , . . . ,m}, >- is the ordering of alternatives I
I
by the criterion i, >O n X i. In the game terms, the problem may be N
i
characterized as follows: the feasible coalitions are the coalitions {i}, i £ N, and the large coalition N; only the latter can efficiently choose an alternative, whereas the choices of coalitions {i} do not affect the result of chance. A planner, designer, governing body, and the like act as the coalition N in practical problems. We consider the problem of finding Pareto-optimal alternatives for functions f i ( x ) , i 6 N on the set of Euclidean space A". We set forth two results obtained by S. Karlin 9 and J.B. Germeyer.6 Lemma 46.1: If A' is convex, /,(x) are concave, and x° is Pareto-optimal,
then there are such AJ, Aj > 0, ^ A, = 1 tnat i
Methodology of Vector Optimization
max J2 Wiiz)
= Yl A*/<(*°) ■
i
419
(46.16) (46-16)
i
Lemma 46.2: If for some A, > 0, i g JV 0 ^A,/,(; a xJKh{x), VA1/,(I), Y,^h^°) = m^Y t ) =
(46.17) (46.17)
iI
then x° is Pareto-optimal. For the case where X is a convex polyhedron, and the functions ft are linear, in Lemma 1 it may be stated 25 that all A; > 0 and for Paretooptimality of the alternative x° it is necessary and sufficient that there would be such Xt with which relation (46.17) takes place. The above lemmas imply that the choice of a particular point in a Pareto-optimaJ set is equivalent, in a sense, to the indication of weights for each criterion. This statement underlies the majority of computational methods. It is also useful for understanding many heuristic methods of defining the optimum with the aid of weights. Lemma 46.3: If x° is Pareto-optimal, here fi(x°) > 0 for all i g N, then there are such A*, A* > 0, i 6 N, J^ Aj = 1, that i
maxmin A \ifi{x) = min minA^/j^ A i / ^ x 00 )).. 2 /i(x) = xEX
i
i
(46.18)
We consider the case, where, apart from the set of alternatives X and objective functions / i ( x ) , . . . , fn(x) we specify some requirements for optimality further called axioms (of optimality). For application, we may need quite different axioms. However, literature still has few axiomatically de fined solutions. Notwithstanding considerable differences in interpretation of the models, for which the systems of axioms were constructed, the axioms themselves are still similar. In all systems we find the variants of axiom symmetry, Pareto-optimality, covariance. Consider the system of axioms of G. Nash and K. Arrow, L. Gurwitz. 18 - 28 G. Nash initially constructed his solution for the problem in transactions concluded by two individuals. Two-individual restriction is non-essential since the Nash axioms are readily generalized for any number of partici pants. But, in order to apply them to the above model, it must be supple mented with one parameter — status quo point (s.q.).
420
Systems Optimization Methodology
In the transaction problem, the s.q. point, which is denoted in terms of /* = (/!*,...,/*), characterizes the state occurring before the beginning of negotiations, and thereby in the case where no agreement is achieved. It is common knowledge that the unambiguous choice of the s.q. point is very hard to obtain. It is also difficult to get a sound answer as to its choice in the present case. It seems natural to choose /* = min/ ; (x), though other iex
variants more suitable for a specific problem are also possible. The decision problem is described by the collection (X, (fl},/*). Let /° = ( / j ° , . . . , f ° ) be the optimal values of objective functions of this model. Consider the formulations of axioms to be satisfied. Axiom 1: (Linear transformation). Let L = (L\,...,Ln) be any linear positive transformation of the function /, then the optimal solution for the
model (A', {Z/./J, L/*} is Lf° In other words, if we change the scale and reference unit of each criterion with respect to the s.q. point, then the optimal values of objective functions change in the same way too. The axiom does not affect behavioral rationality, but merely postulates, in a sense, independence of the optimal solution from choosing the numerical values of utility. Therefore, it may be concluded that it is accomplished in all multiobjective optimization problems. Axiom 2: (Pareto optimality). The optimal value of /° should satisfy the inequality /° ^ /*, i.e. /° > /* for all i € N, f ° must be attainable, i.e. f° — f(x) for a particular x G X, and, finally, there should not be such
x e X that /° <; /(x), i.e. /° < fi(x), for all i € N and f° < f t(x) for some of them. Axiom 3: (Insignificant alternative independence). Let (X, {/,},/*) and
(X1, {/,},/*) be two models, and X D X' If /° is optimal in the first model and /° = /(x) for a certain x 6 X', then f ° is optimal for the second model, too. Accomplishment of this axiom should be carefully examined in each specific problem since it is possible to imagine economic situations, wherein restriction of choice opportunities results in drastic revision of values, optimal "in good life" though achievable in new circumstances, and it is not as required by the axiom. However, if the alternatives are actually not associated with the remaining ones and are insignificant with respect to both models, then the axiom seems to be satisfied.
Methodology of Vector Optimization
421
Axiom 4: (Symmetry). If the model is symmetric, i.e. f* = f* for all i,j € N and for any permutations TT of the set N and x £ X there is such
y e A' that (fi(x)) = (f,.(y)) then /° = /? for all i, j e JV. By interpreting the model under consideration as the transaction problem, we can reformulate the axiom as follows: if the status and opportunities of all negotiators are completely equal, so must be the optimal values of utility obtained by them. Theorem 46.1: /° = f(x) and
The solution /° satisfies axioms 1-4 if and only if
(46.19) To observe mathematical strictness, it should be noted that the Nash solution is given not for an individual model, but for a class of models. The
Nash solution definition domain represents such set of models (X, {/;}, /*) that it may have feasible operations indicated in the axioms: linear positive transformations of the function / and point /* as well as incorporation of the set of values of the function / on X into a symmetric set. Following the Arrow-Gurwitz axiom system we will consider an ar-
bitrary set £7 of objective functions fu(x) on the set of alternatives X. Since the set of objective functions is variable, fi becomes the explicit
component model of decision making. Two models (X,fti,{fu(x)}) and (X, fi 2, {/w(z)}) are referred to as isomorphic if the set constitutes permutation of the set fii. If fa(x) is taken as the matrix with columns u € ft and rows x € X, then the matrices of isomorphic models are derived from one another by rearranging columns. It is assumed that model 2 is derived from model 1 by removing recurrent objective functions if, after removing some uj € fii, model 1 is isomorphic with respect to model 2 and for each removed u> € fii\fi2 there is such w' € fl2 that fu(x) = fu'(x) for all x e X. The class of models P, on which the following axioms and, therefore, the solution are specified, is defined by the following three conditions:
(1) If X is finite and mmfu>(x), maxfu(x) exist for all x e X, then
(x, n, /) € P.
(2) If one model is derived from the other by withdrawing recurrent objective functions and the latter is contained in P, then the model derived must also be contained in P.
422
Systems Optimization Methodology
(3) If the model belongs to P, then for any x 6 X there must be min/ w (ar) UJ
and maxfu(x). UJ
The class P appears to be nonempty. In any event, the problems with a finite number of objective functions fall within this class because the minima and maxima indicated in the definition of P exist for them. Denoting a set of optimal solutions by x°, we state axioms. Axiom 1: If xi C x2 and xi D x% ^ 0. tnen x° = xi C\ x\. This axiom is similar to the Nash axiom of independence from insignificant alternatives differing only in that the Nash solution consists of a unique point whereas this is not needed here. Axiom 2: The sets of optimal alternatives for isomorphic models are the same. In other words, the set of optimal alternatives does not vary with rearrangement of the columns LJ of the matrix ||/w(a:)||-
a(x) < fu(x') for all w € H, then
Axiom 3: If x e x°, x' & X and f x' € x° Similarly, if x 6 x°, x' e X and / w (x) > / w (x') for all w <E ft, then x' £ x° Conceptually, if a particular alternative is not worse by all criteria than an optimal one, then it is optimal, too. If it is not better by all criteria than a certain nonoptimal alternative, then it is also nonoptimal. This condition is utterly indisputable and very feeble as regards application. The result given below is an improvement of the Arrow-Gurwitz theorem. Theorem 46.2: In order that the optimal solution satisfy axioms 1-4 in P, it is necessary and sufficient that (46.20) for all x' e X where > is any ordering of the number pairs (m, M), m < M satisfying the relationship mi > m 2 ,
MI > M2 ==> (mi, MO > (TO 2 ,M 2 ).
(46.21)
In general, the optimal solution defined by axioms 1-3 is constituted by a set rather than an alternative, i.e. it is ambiguous. The solution is ambiguous in that there are several such sets. Each ordering > will correspond to
Methodology of Vector Optimization
423
its set x°, here the more indefinite is the ordering (more equivalent pairs (m, M)) the larger is the set x° The gist of this result is that definition of optimality in the sense of axioms 1-3 for any multicriteria problem amounts to definition of optimality in the two-criteria problem with criteria min/ w (x) and max/ u ,(x). IjJ
UJ
We assume that in place of the objective functions /;(x), the preferences > are specified on X. The group decision is called the rule, following which every collection of preferences (>i)ig;v on X corresponds to the (transitive) preference > N on X. The group decision problem differs from the above problem primarily in that we consider, as a solution, not a set of alternatives, but their ordering. There are also other differences imposed by specific characteristics of application of group decisions to politics. K. Arrow enumerated requirements for the group decision rules: (1) (Universality). The group decision rule is defined for various collections of individual preferences >,, considering that the number of individuals is not less than two, and that of alternatives not less than three. (2) (Positive relation). If for a given set of individual preferences x >N y, then this relation is left also when the individual preferences in pair comparisons including x either do not change or change in favour of x, and other comparisons remain unchanged. (3) (Independence from insignificant alternatives). If all comparisons between alternatives x' C X coincide in two sets of individual preferences, then they have to coincide in group decisions, too. (4) (Sovereignty of citizens). For each pair of alternatives x and y there is such a collection of individual preferences that x >/v y. (5) (Dictator absence). There must not be such individual iQ that x >i y would imply x >N y irrespective of preferences of other individuals. The well-known Arrow paradox resides in the fact that the above conditions are incompatible, i.e. there are no group decision rules satisfying all five conditions. In practice, common use is made of the (simple) majority rule. Simplicity and efficiency of this rule does not need to be proved despite the fact that it sometimes leads to nontransitive social preferences. At present, we know the efficient conditions under which the solution is transitive.1 Moreover, the problem of transitivity arises when there are only two alternatives, i.e. practically in all applications.
424
Systems Optimization
Methodology
The basic features of the simple majority rule are as follows: (1) Definiteness. In any collection of individual orderings, the rule indicates a unique decision for each pair of alternatives. (2) Anonymity. The decision is independent of designations for individuals. (3) Neutrality. The decision is independent of names for alternatives. (4) Positive response. If for a particular set of individual preferences x >N y, and, with the other set, only one individual changes his preference in favour of A" while the rest remains unchanged, then with a new set there should be .r ># VThe meaning of the above conditions for the decision rules is quite evident. It should only be noted that anonymity and neutrality coincide with the corresponding two Arrow-Gurwitz axioms. The following theorem27 characterizes the simple majority rule in full. Theorem 46.3: The simple majority rule represents a unique rule possessing the above features. Let us consider the aggregative criteria problem. Suppose that instead of the alternative orderings >j by the i criteria, we specify the orderings >jt by aggregative criteria k C N. In the general case, we cannot know, of course, the ordering for any k C N, therefore we denote by k the set of those k C N for which the orderings >/t are specified. The problem then is formally described by the collection (A", A", { > k } k e h ' ) - The problem is to find a set of alternatives optimal by a set of criteria {>*:}, or, in a more general case, find the ordering of the set X by this set of criteria. The desired ordering is denoted as >jv. Such problem was first discussed in Ref. 4. The suggested approach can be described conceptually as follows. The orderings are obtained by direct measurements or from individual orderings >;, i € -/V by some aggregation rules. We need to extend these rules to cover the whole society N and thus obtain the ordering >/y. However, we know rather the results of aggregating >k than the rules themselves. Hence, we make it our first business to reveal these rules and find the comparative significance of orderings confined in the aggregative orderings >k- Next, we need to construct such orderings >pj (or their set) in which the revealed relative significance would be preserved and >N fit the most significant aggregative orderings well.
Methodology of Vector Optimization
425
Considering that decision making must proceed from a set of alternatives, i.e. by choosing a point from the set of unimprovable points, in many practical situations it has been found advantageous to employ man-machine procedures. The idea to involve man in the process of decision making was expressed in Ref. 20. The literature describes many procedures of decision making. We will describe some of the most familiar procedures and offer the scheme by which these procedures are constructed. Most of the computational methods for finding the (Pareto) optimal points rely on maximization of the sum on X
Various heuristic procedures share the same idea, too. For example, a researcher is informed about approximate values of the parameters A^ He finds the Pareto-optimal point a:0 corresponding to the particular weights and the rate of change of the function /, depending on the point x° changing in different directions. Proceeding from this information, a decision-maker amends the weights A,, and the problem is solved all over again. B. Roy29 suggested the following general approach to construct the decision-making procedures. The binary relation > on X is called the outranking relation providing that: (1) It is reflexive (x > x for all x e X). (2) For any x, y, z 6 A' the following statements hold true: x > y, /,(j/) >
fi(z) for all f e N => x > z, ft(x) > f1(y) for all t e N, y > z =>• x > z. The outranking relation has the same sense as the above ordering: if x > y, then x is not less preferable than y. Condition
(2) implies (with x = y) that fi(y) > fl(z) for alH € N => y > z, i.e. the outranking relation satisfies the Pareto-optimality axiom if the condition for definition fi(x) > fi(y), i e N is supplemented with the strictness condition even for one inequality and then the consequence is substituted by x > z and z ' f c x , which is quite reasonable. B. Roy presents several specific outranking relations and procedures for their utilization. The general description is as follows. We construct a certain set C C X x X which, by analogy with statistics, may be identified
426
Systems
Optimization
Methodology
as critical. If {x,y) 6 C, then x > y. The procedures rely on construction of the sequence C\ C C^ C • • • Cn, and thereby on construction of the sequence of outranking relations >i C >2 C • • • C >r. The set C may be constructed according to available data in a variety of ways. Consider an example. Let the suggested weights pt of criteria /, be known. Assume that N+(x,y) N+(x,y)
=
N=(x,y)
= {ieN:fl(x)
N-(x,y)
=
{ieN:fl(x)>fl(y)}, =
fl(y)},
{i£N:fl(x)
Consider the functions
P+(x,y) =
£
ft,
i&N+(x,y) ;ejv+(z,y)
P~(x,y)=
Yl
P*' **'
'EN-(x,y)
P=(x,y)=
5Z
Pl
-
i<5N={x,y)
Define the critical set as the set of pairs (x,y) satisfying the inequalities + P+(x,y) P (x,y) + P=(x,y) > >C. C,
P F +(x,y) (.r,y) >
P~(x,y). P-(x,y).
Evidently, the larger is C, the less is the risk (due to the weights p, being not well established) when eliminating alternatives from further consideration, because the advantages of a particular alternative over eliminated ones are more significant. Suppose that, with a certain C\ and > i , alternatives of the set X\ remain uneliminated. If the choice from X\ is hard to make, we have to run the risk of choosing Ci < C\. Repeat calculations and reasonings with Ci in place of C\ and X\ in place of X etc. The method ELECTRA (Elimination and Choice Reflecting the Reality) 15 involves the procedure of multicriteria selection of the "best1' from the set of objects under consideration. For each of the criteria, we introduce a discrete scale of possible values of this criterion (discernable gradations — estimations by this criterion). Thus, each object from the given set E corresponds to the "n-times" Cartesian product of criteria
Methodology of Vector Optimization
427
scales. For each of the criteria, we construct a graph, the vertices of which are separate objects of the set, and the edges point upon a dominance relation among the objects according to the present criterion. Next, we introduce weight coefficients (indices of relative importance) of the adopted criteria. Proceeding from the estimates of objects of the set treated by all criteria with respect to their importance, we calculate the matrices of special coefficient values identified as the indices of agreement or disagreement (to introduce the disagreement index we need first to place in agreement the employed "lengths" of criteria scales.) Further, for each pair of objects (/,/') the outranking relation of, say / over r, is assumed to be established if the value of the corresponding agreement index is larger than a certain threshold value, and the disagreement index is less than the corresponding threshold value. Proceeding from the thus constructed binary relation of outranking, we construct a generalized graph (of outranking), the structure of which essentially depends on the chosen threshold values. The graph kernel contains the best of the available objects as well as those of the objects which are "incomparable" with the chosen thresholds. With the threshold values changing, the number of elements remaining in the kernel is changing, too. The basic advantage of the given approach is likely to reside in the fact that a decision-maker is faced only with the objects, the comparison of which requires a more careful and deeper analysis. Furthermore, variation of threshold values enables one to isolate different collections of objects containing the "best" object and those most unlike it, though featuring the available estimates by the entire totality of criteria does not provide a foundation for these objects to be classed with the deliberately "bad" ones. The possibility of revealing such collections to be considered by the decision-making person seems to be quite reasonable from the standpoint of a more substantiated choice of the "best" object. The method of pairwise comparison of objects is based on the following assumptions: - a decision-maker can always and without difficulty compare two values by any one criteria; - there is the scalar function of utility with continuous first derivatives, by means of which it is possible to carry out complete ordering of all objects being compared. The method involves procedures of successive comparing of object pairs with each other. The most difficult and
428
Systems Optimization Methodology
interesting is the case whereby some portion of criteria object A outranks B, and by another — quite the reverse. The method is discussed at greater length in Ref. 8. A group of methods 8 ' 12 ' 20 has been developed as applied to the problem of linear programming with many criteria. Here an object is taken as the solution belonging to a domain of feasible solutions, due to which the number of objects is not limited and objects are not predetermined. The suggested methods involve a collection of analytical parts, each of which implies presentation of partial results to a decision-maker and utilization of new information obtained from him. The paper of R. Kiney 11 deals with the problem of choosing the best action out of the feasible set, where the results of these actions bear a probabilistic character, i.e. each action features its vector of probabilities for all feasible alternative results. One of the most elaborate theories of choosing the best action in these conditions, generally, is the probabilistic utility theory of von Newman and Morgenstern. By this theory, to evaluate relative preferability of each alternative result under risk we introduce a special utility function, the values of which for each alternative are established on the basis of the decision-maker's suggestions. Based on the functions of each alternative utility and probabilities of their implementation, we form the utility function for each action as a whole. Next we choose the action securing the maximum utility function thus obtained. The alternatives, however, may be distinguished by a number of factors or criteria. Such alternatives have received the name "multidimensional", and each multidimensional alternative, in turn, is evaluated by the multidimensional utility function. Proceeding from the assumption of mutual independence of the factors by utility characterizing the alternatives, R. Kiney suggests the procedure of constructing the multidimensional utility function which demands from a decision-maker actually the same (in practice, minimal possible) body of information as in the additive case, though the field of its application is much wider. The above methods of decision-making with many criteria employ, in one way or another, the additional information obtained from a decisionmaker. In these methods, the stages of carrying out formal mathematical operations (calling for application of computers with large N and n)
Methodology of Vector Optimization
429
alternate with those of dialogues with a decision-maker, which allows the procedures of that type to be identified as man-machine procedures. A new approach to solve the multicriteria optimization problem has been advanced in Ref. 16. The approach is based on the notion of determining an ideal (utopic) point in the space of quality criteria and on introduction of a norm into this space. The tradeoff solution obtained here is the Pareto-optimal solution and secures maximal proximity of quality criteria to their best values. The approach given above is free from choosing weight coefficients, which is difficult especially in solving the large system control problems. A similar idea was almost simultaneously expressed in Ref. 30 for static problems of vector optimization. The same idea was applied in Ref. 17 to solve the optimal design problem for a universal system. For the case of static systems, the above approach has been developed in Refs. 32 and 34. In these works, a suggestion is made to isolate the dominating solutions determined by minimizing the norm more general than the square one and choosing the final solution out of this dominant set proceeding from additional reasonings and information obtained from the person making the final decision. The work in Ref. 32 discusses the properties of tradeoff solution in vector optimization problems. Let X be a solution space, and the space of solution results
Y = (y = f(x) = {/(*!),..., fN(x)}\x e X} is assumed to be compact. Introduce a group regret (for the parameter P)
where P > 1, y* — sup{/j(x)|i £ X}, y £ Y. The point yp of the function
minimum Rp(y) on the set Y and the corresponding point xp = f~l(yp) is referred to as the tradeoff solution for the parameter P. The case P = I leads to maximization of the total utility max^yj; the case P = oo corresponds to min maxfi/* — y, \j'• = 1 , . . . , N}; the case p = 2 corresponds to p j ' the square criterion of M. Salukwadze. It has been proved that the tradeoff solution yp has the following properties: feasibility, minimal group regret,
430
Systems Optimization Methodology
absence of dictator, Pareto-optimality, independence from inconsistent alternatives, symmetry and, in some cases, uniqueness. Further we introduce the concept of a dominance structure which includes the tradeoff solution. Suppose we have two results y1 and y2 in the space Y and d — y2 - y1. If for any A > 0 the result y1 + Xd proves to be preferable than y1, then the vector d is termed the "dominance factor" for the given result y. Let D(y) be a set of all dominance factors for the
given result y supplemented with the null vector in RN (D(y) is a cone). It is assumed that the structure of dominance is specified on Y if for each
y G Y the cone of dominance D(y] is given. Then y1 is dominated by the result y2, if y1 e y2 + D(y2) = {y2 + d\d £ d(y2)}. The result y° is the nondominated solution if there is no result y1 € Y(yi ^ y°) such that
y° £ y1 + D(yl). Evidently, the concept is transferred into the solution space too. The studies in Refs. 33, 34, and the like, have also been carried out in that direction. In conclusion we will discuss the solution optimization problem in hierarchical systems.7 Optimization of solutions in hierarchical systems or with respect to them (synthesis problem) encounters many difficulties, which is determined by specific character of hierarchical systems: distribution by different hierarchical levels, subordination of one level to another, multiplicity of objectives and other conditions for functioning. As a result of diversity of the nature of vector optimization problems, it is not improbable that the above difficulties can be largely obviated by employing the vector optimization formalism. 2 ' 3 Let A' be a hierarchical system, the behavior of which is determined by the parameters: (1) J = { 1 , 2 , . . . , n] — a set of subsystems; (2) {x € A',},6j, x 6 X — n Xi feasible sets of subsystems and the whole ^€J system; (3) {x —> 3/,(x)}igj-, y = ( 3 / 1 , . . . ,7/ n ) the scalar criteria of subsystems and the vector criterion of the system. Conditions 1-3 determine behavior of any system. Introduce additional conditions determining a hierarchical structure of the system; (4) J = {1,2,..., m} is a set of hierarchical levels;
Methodology of Vector Optimization
(5) (6)
is distribution of subsystems by levels; = (XI, X2,. . . ,Am) is a set of binary relations of the priority of levels or the priority vector.
(J,JjEJ
j,X
Behavior of the hierarchical system is determined by six parameters:
on the basis of which the search is carried out for the optimal system X O . We first consider behavior of the system without Conditions 4-6. In this case we have the general problem of vector optimization of systems by the criterion y(x) = (yl,yz, . . . ,y,). Without detriment to Conditions 4-6 it is possible to define the tradeoff domain, owing to which the field of search for an optimal system is restricted. Furthermore, in a number of cases we have t o restrict optimization of the system by this
Model (46.23) can be implemented either accurately or approximately, say, by employing the notion of the directed search or the properties of local optima. In a number of cases, model (46.23) may be employed to evaluate satisfactoriness of the system structure from the condition:
which means that in the systems of a given structure there is no strict rivalry of subsystems, and the structure is feasible. Condition (46.24) may be rather simply tested by evaluating variations in local optima. Further intrusion into the tradeoff domain to search for a unique optimal system should be carried out with compulsory utilization of Conditions 4-6. They allow a general problem (46.22) t o be naturally decomposed into two classes of vector optimization subproblems - those for optimization at each hierarchical level and that of interlevel optimization. We consider subproblems of the first class. At each jth hierarchical level there is the vector optimization problem without priority of the form:
Here y(-l) is a vector criterion for the jth hierarchical level with componentslocal criteria of all subsystems of zs given level.
432
Systems Optimization Methodology
Problem (46.25) is solved only conceptually, i.e. the search for an op-
timum on the feasible set X" = n Xl by the criterion yJ} is not carried
;eJj out. We perform merely a convolution of the vector criterion into a generalized scalar criterion on a basis of tradeoff axiomatics and its corresponding computational scheme
yu)^
or from the "uniformity" axiomatics (46.28) As noted above, the tradeoff axiomatics may be chosen either analytically or heuristically, commonly by employing additional information on the system. As a result of vector criteria convolution, each hierarchical level obtains a generalized scalar criterion ?_, (x), which enables one to go over to solution of the second class subproblems of interlevel optimization
(J, {Xj},ej, te(z)}je>7, {Aj-W) (46.29) The problem refers to the class of vector optimization problems with priority. It should be first solved conceptually, too. Convolution of the vector criterion ip = (<^i,>2, • • • ,>m) is carried out with regard to the priority vector A (V, A) -> w = u(v, A ) . (46.30) In considerable differentiation of priority relations, it is also possible to construct an algorithmic solution of the lexicographic or quasilexicographic ordering type. Thus utilization of vector optimization allows a sound procedure of choosing the optimal solution in hierarchical systems to be constructed.
Methodology of Vector Optimization
433
In this case the complex original optimization problem (46.22) reduces to the trivial scalar optimization problem (or a sequence of such problems). X° =fX° =f-11\maxuj((ip,x),\)]. [maxu{(ip,x),\)
(46.31)
.
To illustrate, in utilizing the axiomatics of the valid concession tradeoff for interlevel optimization and optimization at each hierarchical level there is the scalar problem: 1 X° X° = = frl
Here
m m
m aa xx fj ]] f( D n >^ ( *W) ) K m *i
/I // mm m m
•■
(46.32) (46-32)
\\
In this problem, when constructing the procedure of choosing a solution, account was taken of two features of hierarchical systems — hierarchism and multiplicity of objectives. In this case, on the way to solution (46.31), the formalism of vector optimization was utilized three times: in evaluat ing the entire system as a whole regardless of the hierarchical structure; in defining the decision procedure at each hierarchical level (46.25); for interlevel optimization (46.29). Similarly, account may also be taken of other factors of the given class systems. CONCLUSION Starting in the middle of the 60s, the trend was integration of different specialized and sufficiently developed theories with the object of founding a more abstract theory for a particular class of systems. However, due to the fact that the formal apparatus of such theory requires a fairly rich mathematical structure for adequate description of large-scale or purposeful systems, we may presently discuss essential alterations in the available for mal apparatus rather than development of the new one capable of rising up to the level of generalization needed here. In our opinion, proceeding from unified positions and relying basically on a unified mathematical structure employed in systems description, a new formal apparatus is to investigate, in the hope of obtaining conceptual results, such different problems as that
434
Systems Optimization Methodology
of existence and minimality for a set of axioms providing an opportunity for representations in the state space, the necessary and sufficient conditions for controllability and optimization of multidimensional systems, the problem of minimal implementation of regularities relating input actions to output values, the necessary and sufficient conditions for stability in a wide range of understanding of this category. To be investigated is the Godel consistency and completeness theorem, the multimeasure system non-interaction problem, the Krione-Rodes decomposition theorem and the system classification problem based on utilization of the apparatus of a new, formal systems theory. The present monograph touched upon only the most fundamental concepts and aspects of the systems investigation and optimization. Many theories of specific type systems have been developed for a fairly extended time. We attempted to consider the problems of analysis and synthesis and those of the systems multiobjective optimization, taking them as basic and common to more specialized disciplines. Of vital importance is the possibility of studying the behavior of a system, investigating the decision-making processes proceeding in this system or the mechanisms ensuring the purposefulness of its behavior. Research in the problems of the systems optimization methodology may be beneficial in investigating the systems under uncertainty where information on the system and regularities of its functioning proves to be incomplete and insufficient for constructing its detailed mathematical model. In the case of investigating large-scale and complex systems, it is possible to raise the efficiency of analysis and synthesis of the system behavior or just to provide such possibility if the system optimization methodology be employed, admitting development of a less structured model which merely utilizes key factors. In general, structural considerations have a dominant role both in analysis and in synthesis of the different types of systems. The systems optimization methodology may serve as a language of interdisciplinary exchange since it is sufficiently general for not introducing its constraints and to be considered, at the same time, as a unified foundation for more narrow directions of systems investigation.
REFERENCES References
to Chapter 2
1. Bacon R., "Large Writing", Anthology of World Philosophy, Moscow, Misl, 1969, vol. I, part 2, p. 865. 2. Anthology of World Philosophy, Moscow, Misl, 1969-1972, vol. I-IV, vol. 3, pp. 313-314. 3. Vindelband V., History and Natural History: Preludes: Philosophical Articles and Discourses, St. Petersburg, 1904. 4. Rickert H., Sciences of Nature and Sciences of Culture, St. Petersburg, 1911. 5. Kant Philosophy and Modern Times, Moscow, Misl, 1974.
6. Von Neumann J. and O. Morgenstern, Theory of Games and Economic Behavior, Moscow, Nauka, 1970. 7. Suppes P. and G. Zines, Fundamentals of Measurement Theory. Psychology of Measurements, Moscow, Mir, 1970.
8. Tarski A., Logic, Semantics and Metamathematics, Oxford University Press, 1956, p. 418. 9. Taube M., Computers and Common Sense, Moscow, Progress, 1964, p. 110.
References to Chapters 3—7 1. Allen R., Mathematical Economics, Foreign Literature Publishing House, 1963. 2. Ben-Israel A. and P.D. Roberts, "A decomposition method for interval linear programming", Manage. Sci. 16, 5 (1970). 3. Bulavsky V.A., R.A. Zviagina, and M.A. Yakovleva, Numerical Methods for Linear Programming, Moscow, Nauka, 1977. 4. Vasilyev P.P., Numerical Methods for Solving Extremal Problems, Moscow, Nauka, 1980.
5. Vatel LA. and J.A. Fliorov, "A model for annual planning in an industry", in Program Method of Control, Moscow, C.C. in the USSR Academy of Sciences, part 3, 1976. 435
436
Systems Optimization Methodology
6. Bakhutinsky I.J., L.M. Dudkin, and B.A. Shchennikov, "Iterative aggregation in some optima] economic models'', Economics and Mathematical Methods IX, 3 (1973). 7. Verina L.F. and V.S. Tanayev, "Decomposition approaches in solving mathematical programming problems", Economics and Mathematical Methods XI, 6 (1975). 8. Williams A.C., "A treatment of transportation problem by decomposition", SIAM J. Appl. Math. 10, 1 (1962). 9. Goldstein J.G. and D.B. Yudin, New Trends m Linear Programming, Moscow, Sov. Radio, 1966. 10. Goldstein I.G., "On the possibility of expanding applicability of special methods for linear programming", in Planning and Economic-Mathematical Methods, Moscow, Nauka, 1971. 11. Dantzig G., Linear Programming, its Generalizations and Application, Moscow, Progress, 1966. 12. Yermolyev J.M. and L.G. Yermolyeva, "Parametric decomposition method", Cybernetics 2 (1973). 13. Yermolyev J.M., "Methods for solving non-linear extremal problems'', Cybernetics 4 (1966). 14. Cornai I. and T. Liptack, "Two-level planning" in Application of mathematics in economic studies, vol. 3, Moscow, Misl, 1965. 15. Kutikov L.M., "Decomposition of linear programming block problems with weakly associated blocks", Economics and Mathematical Methods IX, 4 (1973). 16. Kutikov L.M. and S.A. Gesan, "Simplex method for solving linear dynamic programming problems", Economics and Mathematical Methods XVI, 6 (1980). 17. Lebedev V.J., "Decomposition method for solving the linear programming block problems with the coupling variables", Journal Computational Mathematics and Mathematical Physics 21, 4 (1981). 18. Lebedev V.J., "Scheme for solving partially integer problems of mathematical programming", Journal Computational Mathematics and Mathematical Physics 17, 5 (1977). 19. Lesdon L.S., Optimization of large systems, Moscow, Nauka, 1975. 20. Malinnikov V.V., "Method for expansion in solving large linear programming problems with a block structure", Economics and Mathematical Methods VII, 5 (1971). 21. Mauer I., "Penalty constant in block programming", Proc. of the Estonian Acad. of Sciences, Physics, and Mathematics, vol. 20, no. 4, 1971. 22. Mauer I., "On utilization of penalty constant in decomposition of mathematical programming problems", Proc. of the Estonian Acad. of Sciences, Physics, and Mathematics, vol. 20, no. 4, 1971.
References 437
23. Mednitsky V.G., "On optimality of aggregation in the linear programming block problem" in Mathematical Methods for Solving Economic Problems, part 3, 1972. 24. Meressoo T.H., "On one decomposition method for the resources allocation problem", Proc. of the Estonian Acad. of Sciences, Physics, and Mathematics, vol. 27, no. 2, 1978. 25. Mesarovich M., D. Mako, and I. Takahara, Theory of Hierarchical Multilevel
Systems, Moscow, Mir, 1973. 26. Mikhalevich V.S., J.M. Yermolyev, V.V. Shkurba, and N.Z. Shor, "Complex systems and solutions of extremal problems1', Cybernetics 5 (1967). 27. Moiseyev N.N., J.P. Ivanilov, and J.M. Stolyarova, Optimization Methods, Moscow, Nauka, 1978. 28. Pervozvanskaya T.N. and A.A. Pervozvansky, "The algorithm of search for the optimal centralized resources allocation", Proc. of the USSR Academy of Sciences, Engineering Cybernetics, no. 3, 1966. 29. Pervozvanskaya T.N., "Optimal planning in the linear dynamic model with regard to uncertainty", in Problems of Applying Mathematics in the Socialist Economy, no. 2, Leningrad University, 1965. 30. Pervozvanskaya A.A. and V.G. Gaytzgori, Decomposition, Aggregation and Approximate Optimization, Moscow, Nauka, 1979. 31. Pshenichny B.N. and J.M. Danilin, Numerical Methods in Extremal Problems, Moscow, Nauka, 1975. 32. Ritter K., "A decomposition method for linear programming problems with coupling constraints and variables", Mathematics Research Center, University of Wisconsin, Rep. 739, 1967. 33. Rosen J.B., "Primal partition programming for block diagonal matrices", Numerische Mathematik 6 (1964). 34. Sekine J., "Decentralized optimization of an interconnected system", IEEE Trans. Circuit Theory, June (1963). 35. Ulm S., "Componentwise descent and hierarchical optimization", Proc. of the Estonian Academy of Sciences, Physics, Mathematics, vol. 18. no. 1, 1969. 36. Ulm S., "Conditional gradient method and decomposition of optimal control problem", Proc. of the Estonian Academy of Sciences, Physics, Mathematics, vol. 18, no. 1, 1969. 37. Ulm S., "On decomposition of the Resources allocation problem based on the interactions prediction principle", in Models and Methods for Analysis of Economic Purposeful Systems, Novosibirsk, Nauka, 1977. 38. Ulm S. and T.H. Meressoo, "On decomposition of the resources allocation type problem1', in Methods for Interaction Analysis in Economic Systems, Novosibirsk, Nauka, 1980. 39. Hadley G., Linear Programming, Addison-Wesley Publ. Co., Reading, Mass., 1962.
438
Systems Optimization Methodology
40. Tsurkov V.I., Decomposition in Large Dimension Problems, Moscow, Nauka, 1981.
41. Shor N.Z., "Application of the generalized gradient descent in block programming", Cybernetics 3 (1967). 42. Shor N.Z., "On convergence of the generalized gradient method", Cybernetics 3 (1968). 43. Shchennikov B.A., "Block method for solving a system of large dimension linear equations", Economics and Mathematical Methods I, 6 (1965). 44. Shchennikov B. A., "Application of the iterative aggregation method for solving systems of linear equations", Economics and Mathematical Methods II, 5 (1966). 45. Shchennikov B.A., "Aggregation method for solving systems of linear equations", Latvian Acad. of Sci. of USSR 173, 4 (1967). References to Chapter 8 1. Rohlich, Breianer Das Getreide, Verlag A.W. Hayni Arben, Berlin 5036. 2. Preidrich H.P., Der Endproduzent in der Getrlide-noivtschaft, Leitschrift Agroforum 6/7, 1968. 3. Kolbin V.V., Macromodels of the National Economy of the USSR, D. Reidel Publishing Company, 1985. 4. Statisticher Yahrbuch der VEAB und KFM, 1967. 5. Walter Ulbricht: Rede auf dem VII Parteitag der SED. 6. Yudin D.B. and J.G. Goldstein, Linear Programming. Theory and Finite Methods, M., 1963.
References to Chapter 9 1. Arrow K.I., "Rational choice functions and orderings' , Econometrica 26 (1959) 121-127. 2. Arrow K.J., Social choice and individual values, 2nd edn., Yale Univ. Press, New Haven, 1970, p. 124. 3. Arrow K.J. and L. Hurwitz, "An optimality criterion for decision-making under ignorance'', in Uncertainty and Expectation in Economics, ed. by C. F. Carter and J.L. Ford, Oxford: Blackwell and Mott, 1972, p. 299. 4. Khachiyan L.G., "Convexity and complexity of solving polynomial programming problems", Proc. of the USSR Academy of Sciences, Engineering Cybernetics, no. 6, 1982. 5. Khachiyan L.G., "Complexity of convex problems for integer polynomial programming", Auther's Abstract of Dissertation, Moscow, All-Union Central Information Agency, USSR, 1984.
6. Khomenyuk V.V., Optimal Control Systems, Moscow, Nauka, 1977, p. 152. 7. Khomenyuk V.V., Elements of the Multiobjective Optimization Theory, Moscow, Nauka, 1983, p. 123.
References 439 8. Kolbin V. V., Methods for Multiobjective Optimization, Ail-Union Institute of Scientific and Technical Information, 1987. 9. Milnor J., "Games against nature", in Decision Processes, ed. by R.M. Thrall, C.H. Coombs, and R.L. Davis, Wiley; L.: Chapman and Hall, 1954, p. 331. 10. Nash J.F., "The bargaining problem", Econometrica 18 (1950) 155-162. 11. Podinovsky V.V., "On the relative importance of criteria in multicriteria decision problems", in Multicriteria Decision Problems, Moscow, Mashinostroyeniye, 1978. 12. Sen A.K., Collective Choice and Social Welfare, San Francisco: Holden Day; Edinburg; London; Oliver and Boyd, 1970, p. 225. 13. Sen A.K., "Choice functions and revealed preference", Rev. Econ. Stud. 38 (1971) 307-317.
References to §44 1. Benayun R., O.I. Larichev, G. Mongolfye, and G. Terny, "Linear programming with multiple criteria", Automatics and Telemety, 8 (1971) 108-115. 2. Ventzel J.S., "Introduction to operations research", Moscow, Sov. Radio, 1964. 3. Vilkas E., "Theory of utility and decision-making", in Mathematical Methods in Social Sciences, no. 7, Vilnus, 1971, pp. 13-60. 4. Vilkas E., "Multiobjective optimization", in Mathematical Methods in Social Sciences, no. 7, Vilnus, 1976, pp. 16-67. 5. Germeyer J.B., Introduction to Theory of Operations Research, Moscow, Fizmatgiz, 1971. 6. Diner I.Y., "Zoning the nature state vector set and the decision choice problem-Problems of general and social prediction", no. 2. Operations Res. Information Bulletin, Moscow, 1969, no. 14.29 7. Dubrov J.A., L.A. Grechina, and V.T. Kulik, "On decision making under uncertainty (systems approach)", Trans. 7th Symp. on Cybernetics, Tbilisi,
1972. 8. Yemelyanov S.V., V.I. Borisov, A.A. Malevich, and A.M. Cherkashin, "Models and methods of vector optimization1' - Results of Science and Engineering, Engineering Cybernetics, Vol. 5, Moscow, 1973. 9. Yemelyanov S.V., J.B. Dudin, O.I. Larichev, A.A. Malevich, E.L. Nappelbaum, and V.M. Ozernoy, "Preparation and decision making in the organizing control systems'', Results of Science and Engineering, Engineering Cybernetics, Moscow, 1971. 10. Karlin S., Mathematical Methods in the Game Theory, Programming and Economics, Moscow, Mir, 1964. 11. Lebedev B.D., V.V. Podnovsky, and R.S. Styrikovich, "Optimization problem for ordering a set of criteria" , Economics and Mathematical Methods, Vol. VII, no, 4, 1971.
440
Systems Optimization Methodology
12. Lewis R. and X. Raypha, Games and Solutions, Moscow, FL., 1961. 13. Podinovsky V.V. and V.M. Gavrilov, Optimization by Successively Applied Criteria, Moscow, Sov. Radio, 1975. 14. Podinovsky V.V., "Multicriteria problems with homogeneous criteria ordered
by importance'', Automatics and Telemetry, no. 1, 1976. 15. Salukvadze M.J., Problems of Vector Optimization in Control Theory, Tbilisi, Metsniereba, 1975. 16. Salukvadze M.J., "Optimal control problems with several quality criteria'', Author's Abstract of the Dissertation, Moscow, 1974. 17. Trukhayev R.I., Research Methods for Decision Making under Uncertainty, Leningrad, 1972. 18. Federov V.V., Numerical Maximin Methods, Moscow, Nauka, 1979. 19. Von Nauman G. and N.O. Morgenstern, Theory of Games and Economic Behavior, Moscow, Nauka, 1970. 20. Debrez G., Theory of Value, John Wiley, New York, 1959. 21. Dong Hak Chyung, "Optimal systems with multiple cost functionals", J. SIAM Control 5, 3 (1967). 22. Kuhn H.W. and A.W. Tucker, "Nonlinear programming' 1 , Proc. Second Berkeley Symp. on Mathematic Statistics and Probability, LJniversity of California Press, Berkeley, California, 1951. 23. Nelson W.L., "On the use of optimization theory for practical control system design", IEEE Trans. Automatic Control AC-9, 4 (1964). 24. Pareto V., Cour's d'Economie Politique, Lausanne. Rouge, 1896. 25. Waltz F.M., "An engineering approach: Hierarchical optimization criteria", IEEE Trans. Automatic Control AC-12, 2 (1967). 26. Zadeh L.A., "Optimality and non-scalar valued performance criteria", IEEE Trans. Automatic Control AC-8, 1 (1963). References to §^5 1. Gorokhovik V.V., "To the vector optimization problem", Proc. of the USSR Acad. of Science, Engineering Cybernetics, no. 6, 1972. 2. Yemelyanov S.V., V.I. Borisov, A.A. Malevich, and A.M. Cherkishin, "Models and methods for vector optimization", Results of Science and Engineering, Engineering Cybernetics, vol. 5, 1973. 3. Von Neuman G. and O. Morgenstern, Theory of Games and Economic Behavior, Moscow, Nauka, 1970. 4. Salukvadze M.J., "Problems of vector optimization in control theory", Metsniereba, Tbilisi, 1975. 5. Bragaard L. and J. Vangeldere, "Points efficaces en programmation a objectives multiples", Bull. Soc. Roy. Sci. Liege 46, 1-2 (1977) 27-41. 6. Chang and Sheldon S.L., "General theory of optimal process", J. SIAM Control 4, 1 (1966).
References 441 7. Da Cunha N.O. and E. Polak, "Constrained minimization under vectorvalued criteria in linear topological spaces", Mathematical Theory of Control, Academic Press, 1967. 8. Ecker J.G. and J.A. Kanada, "Finding efficient points for linear multiple objective programs", Math. Programs (1975) 375-377. 9. Klinger A., "Vector valued performance criteria", IEEE Trans. Automatic Control AC-8, 1 (1963). 10. Koopmans T.C., Analysis of Production. An Efficient Combating of Activities. Activity Analysis of Production, New York, 1951. 11. Reid R. and S.J. Citron, "On noninferior performance index vectors", J. Optimization Theory and Application 7, 1 (1971). 12. Stern Ronald J. and Adi. Ben-Israel, "An interior penalty function method for the construction of efficient points in a multicriteria control problem", J. Math. Anal. Appl. 46, 3 (1974) 768-776. 13. Steuer Ralph E., "Multiple objective linear programming with interval criterion weights", Manage. Sci. 23, 3 (1976) 305-316. 14. Steuer Ralph E., "A five phase procedure for implementing a vector-maximum algorithm for multiple objective linear programming problems1', Lect. Notes Econ. and Math. Syst., (1976). 15. Wendell R.E. and D.N. Lee, "Efficiency in multiple objective optimization problems", Math. Program 12, 3 (1977) 406-414. 16. Zadeh L.A., "Optimality and non-scalar valued performance criteria", IEEE Trans. Automatic Control. AC-8, 1 (1963).
References to §^# 1. Apinis A., "Transitivity of group decisions according to the simple majority rule", Lithuanian Mathematical Problems, vol. 14, no. 4, 1974. 2. Borisov V.I., "Vector optimization of systems", in Systems Research, Moscow, 1971. 3. Borisov V.I., "Vector optimization solutions in organizing systems." Heads of reports at the Int. Symp. on the Organization Control Problems and Hierarchical Systems, Baku, Moscow, 1971. 4. Vilkas E., "Co-ordination of coalition orderings'', in Achievements of the Game Theory, Mintis, Vilnus, 1973, pp. 83-92. 5. Vilkas E., "Multiobjective optimization", in Mathematical Methods in Social Sciences, no. 7, Vilnus, 1976, pp. 16—67. 6. Germeyer J.B., Introduction to Operations Research, Nauka, Moscow, 1971. 7. Yemelyanov S.V., V.I. Borisov, A.A. Malevich, and A.M. Cherkishin, "Models and methods for vector optimization", Summary of Science and Engineering, Engineering Cybernetics, vol. 5, 1973.
442
Systems Optimization Methodology
8. Yemelyanov S.V., V.M. Ozernoy, and O.I. Larichev, "Problems and methods of decision making 1 ', International Centre of Scientific and Technical Information, Moscow, 1973. 9. Karlin S., Mathematical Methods m the Game Theory, Programming and Economics, Mir, 1964. 10. Kafarov V.V., G.B. Lazarev, and V.I. Avdeyev, "The method of solving the multicriteria control problems for a complex chemical-technological system", USSR Acad. of Sci., vol. 198, no. 1, 1971. 11. Kiny R., "Utility functions of multidimensional alternatives", in Problems of Analysis and Procedures of Decision Making, Mir, Moscow, 1976. 12. Larichev O.I., "Man-machine procedures of decision making", Automatics and Telemetry, no. 12, 1971. 13. Lews R. and X. Raifa, Games and Decisions, Moscow, Foreign Literature, 1961. 14. Von Neuman G., Morgenstern O., Theory of Games and Economic Behavior, Moscow, Nauka, 1970. 15. Roy B., "Classification and choice with several criteria (ELECTRA method)" , in Problems of Analysis and Procedures of Decision Making, Mir, Moscow, 1976. 16. Salukvadze M.J., Problems of Vector Optimization in Control Theory, Metsniereba, Tbilisi, 1975. 17. Tokarev V.V., "Optimization of parameters of the dynamic system universal for a series of manoeuvres with different degrees of information", Automatics and Telemetry, no. 8, 1971. 18. Arrow K.J. and L. Hurwitz, "An optimality criterion for decision-making under ignorance", in Uncertainty and Expectation in Economics, Basil Blackwell and Mot, Oxford, 1972. 19. Arrow K . J . , Social Choice and Individual Values, Cowels Comission Monograph, 12, Wiley, N.Y., 1951. 20. Benayoun R. and J. Tergny, "Cnteres multiples et programmation mathematique: une solution dans le cas linear", R.I.R.O., no. V-2, 1969. 21. Bergstresser K., A. Charnes, and P.L. Yu, "Generalization of domination structures and nondominated solutions in multicriteria decision making", J. Optim. Theory Appl. 18, 1 (1976) 3-13. 22. Black D., The Theory of Committees and Elections, Cambridge University Press, 1958. 2.3. Bogaard P.J.M. and J. van den Versluis, "The design of optimal committee decision'', Statistical Neerlandica, v. 16, 1962. 24. Fishburn P.C., "Additive utilities with incomplete product sets: Application to priorities and assignments", Oper. Res. 15, 3 (1967). 25. Focke J., "Vector maximum problem und parametrische optimierung' , Math. Operations forsch. und Statist. 4, (1973) Heft 5, 365-369.
References 443
26. Harsanyi J.C., "Cardinal welfare. Individualistic ethics and interpersonal comparisons of utility", J. Political Economy 63 (1955). 27. MAYKO, "A set of independent necessary and sufficient conditions for simple majority decision", Econometnca 20 (1952) 680-684. 28. Nash J., "The bargaining problem", Econometnca 18 (1950) 155-162. 29. Roy B., "How outranking relation helps multiple criteria decision making: METRA", Rapport de Recherche, no. 67, Sept., (1972). 30. Salama A.J. and V. Gurishankar, "Optimal control of systems with a single control and several cost functionals", Int. J. Control 14, 4 (1971). 31. Shlaifer R.O., Analysis of Decision under Uncertainty, McGraw-Hill, 1967.
32. Yu P.L. and G. Leitmann, "Compromise solutions, domination structures and Sulakvadse's solution", MuHicriieria Decision Making and Differential Games, 1976, pp. 85-101. 33. Yu P.L. and G. Leitmann, "Nondominated decision and cone convexity in dynamic multicriteria decision problems", Multicriteria Decision Making and Differential Games, 1967, pp. 61-72. 34. Yu P.L., "Cone convexity, cone extreme points and nondominated solutions in decision problems with multiobjectives", Multicriteria Decision Making and Differential Games, 1976, pp. 1—59.
35. Zadeh L.A., "Optimality and non-scalar valued performance criteria", IEEE. Trans. Automatic Control AC-8, 1 (1963).
Sign In