Professional Documents
Culture Documents
Robert Fullér
Department of Operations Research
Eötvös Loránd University
Pázmány Péter sétány 1C,
P. O. Box 120, H-1518 Budapest, Hungary
e-mail: rfuller@cs.elte.hu
and
Institute for Advanced Management Systems Research
Åbo Akademi University,
Lemminkäinengatan 14B, Åbo, Finland
e-mail: rfuller@mail.abo.fi
Péter Majlender
Turku Centre for Computer Sceince
Institute for Advanced Management Systems Research
Åbo Akademi University,
Lemminkäinengatan 14B, Åbo, Finland
e-mail: peter.majlender@mail.abo.fi
One important issue in the theory of Ordered Weighted Averaging (OWA) oper-
ators is the determination of the associated weights. One of the first approaches,
suggested by O’Hagan, determines a special class of OWA operators having max-
imal entropy of the OWA weights for a given level of orness; algorithmically it is
based on the solution of a constrained optimization problem. Another considera-
tion that may be of interest to a decision maker involves the variability associated
with a weighting vector. In particular, a decision maker may desire low variability
associated with a chosen weighting vector. In this paper, using the Kuhn-Tucker
second-order sufficiency conditions for optimality, we shall analytically derive the
minimal variablity weighting vector for any level of orness.
Keywords: Multiple criteria analysis, Fuzzy sets, OWA operator, Lagrange multi-
plier
1 Introduction
The process of information aggregation appears in many applications related to
the development of intelligent systems. One sees aggregation in neural networks,
fuzzy logic controllers, vision systems, expert systems and multi-criteria decision
aids. In [6] Yager introduced a new aggregation technique based on the ordered
weighted averaging operators.
Definition 1.1. An OWA operator of dimension n is a mapping F : Rn → R that
has an associated weighting vector W = (w1 , . . . , wn )T of having the properties
w1 + · · · + wn = 1, 0 ≤ wi ≤ 1, i = 1, . . . , n, and such that
n
F (a1 , . . . , an ) = wi bi ,
i=1
where bj is the jth largest element of the collection of the aggregated objects
{a1 , . . . , an }.
A fundamental aspect of this operator is the re-ordering step, in particular an
aggregate ai is not associated with a particular weight wi but rather a weight is
associated with a particular ordered position of aggregate. When we view the OWA
weights as a column vector we shall find it convenient to refer to the weights with
the low indices as weights at the top and those with the higher indices with weights
at the bottom. It is noted that different OWA operators are distinguished by their
weighting function.
In [6], Yager introduced a measure of orness associated with the weighting
vector W of an OWA operator, defined as
n
n−i
orness(W ) = · wi .
n−1
i=1
and it characterizes the degree to which the aggregation is like an or operation. It
is clear that orness(W ) ∈ [0, 1] holds for any weighting vector.
It is clear that the actual type of aggregation performed by an OWA operator
depends upon the form of the weighting vector [9]. A number of approaches have
been suggested for obtaining the associated weights, i.e., quantifier guided aggre-
gation [6, 7], exponential smoothing [3] and learning [10].
Another approach, suggested by O’Hagan [5], determines a special class of
OWA operators having maximal entropy of the OWA weights for a given level
of orness. This approach is based on the solution of he following mathematical
programming problem:
n
maximize disp(W ) = − wi ln wi
i=1
n
n−i
subject to orness(W ) = · wi = α, 0 ≤ α ≤ 1 (1)
n−1
i=1
w1 + · · · + wn = 1, 0 ≤ wi , i = 1, . . . , n.
1
In [4] we transfered problem (1) to a polinomial equation which is then was
solved to determine the maximal entropy OWA weights.
Another interesting question is to determine the minimal variablity weighting
vector under given level of orness [8]. We shall meausure the variance of a given
weighting vector as
n
1
D2 (W ) = · (wi − E(W ))2
n
i=1
n
1 n n
2
1 2 1 1
= wi − wi = wi2 − 2 .
n n n n
i=1 i=1 i=1
1 1
D2 (1, 0, . . . , 0, 0) = D2 (0, 0, . . . , 0, 1) = − 2.
n n
n n
n
1 1 n−i
L(W, λ1 , λ2 ) = wi2 − 2 + λ1 wi − α + λ2 wi − 1 .
n n n−1
i=1 i=1 i=1
2
denote the Lagrange function of constrained optimization problem (2), where
λ1 and λ2 are real numbers. Then the partial derivatives of L are computed as
∂L 2wj n−j
= + λ1 + · λ2 = 0, 1 ≤ j ≤ n, (3)
∂wj n n−1
n
∂L
= wi − 1 = 0,
∂λ1
i=1
n
∂L n−i
= · wi − α = 0.
∂λ2 n−1
i=1
We shall suppose that the optimal weighting vector has the following form
∂L 2wp n−p
= + λ1 + · λ2 = 0
∂wp n n−1
and for j = q we get
∂L 2wq n−q
= + λ1 + · λ2 = 0
∂wq n n−1
That is,
2(wp − wq ) q − p
+ · λ2 = 0
n n−1
and therefore, the optimal values of λ1 and λ2 (denoted by λ∗1 and λ∗2 ) should
satisfy the following equations
∗ 2 n−q n−p n−1 2
λ1 = · wp − · wq and λ∗2 = · · (wq − wp ). (5)
n q−p q−p q−p n
Substituting λ∗1 for λ1 and λ∗2 for λ2 in (3) we get
2 2 n−q n−p n−j n−1 2
· wj + · wp − · wq + · · · (wq − wp ) = 0.
n n q−p q−p n−1 q−p n
That is the jth optimal weight should satisfy the equation
q−j j−p
wj∗ = · wp + · wq , j ∈ I{p,q} . (6)
q−p q−p
3
From representation (4) we get
q
wi∗ = 1,
i=p
That is,
q
q−i i−p
· wp + · wq = 1,
q−p q−p
i=p
i.e.
2
wp + wq =.
q−p+1
From the constraint orness(w) = α we find
q q q
n−i n−i q−i n−i i−p
· wi = · · wp + · · wq = α
n−1 n−1 q−p n−1 q−p
i=p i=p i=p
that is,
2(2q + p − 2) − 6(n − 1)(1 − α)
wp∗ = . (7)
(q − p + 1)(q − p + 2)
and
2 6(n − 1)(1 − α) − 2(q + 2p − 4)
wq∗ = − wp∗ = . (8)
q−p+1 (q − p + 1)(q − p + 2)
The optimal weighting vector
W ∗ = (0, . . . , 0, wp∗ , . . . , wq∗ , 0 . . . , 0)T
is feasible if and only if wp∗ , wq∗ ∈ [0, 1], because according to (6) any other wj∗ ,
j ∈ I{p,q} is computed as their convex linear combination.
Using formulas (7) and (8) we find
1 2q + p − 2 1 q + 2p − 4
wp∗ , wq∗ ∈ [0, 1] ⇐⇒ α ∈ 1 − · ,1 − ·
3 n−1 3 n−1
The following (disjunctive) partition of the unit interval (0, 1) will be crucial in
finding an optimal solution to problem (2):
n−1
n−1
(0, 1) = Jr,n ∪ J1,n ∪ J1,s . (9)
r=2 s=2
where
1 2n + r − 2 1 2n + r − 3
Jr,n = 1− · , 1− · , r = 2, . . . , n − 1,
3 n−1 3 n−1
1 2n − 1 1 n−2
J1,n = 1− · , 1− · ,
3 n−1 3 n−1
1 s−1 1 s−2
J1,s = 1− · , 1− · , s = 2, . . . , n − 1.
3 n−1 3 n−1
4
Consider again problem (2) and suppose that α ∈ Jr,s for some r and s from
partition (9). Such r and s always exist for any α ∈ (0, 1), furthermore, r = 1 or
s = n should hold.
Then
W ∗ = (0, . . . , 0, wr∗ , . . . , ws∗ , 0 . . . , 0)T (10)
where
wj∗ = 0, if j ∈
/ I{r,s} ,
2(2s + r − 2) − 6(n − 1)(1 − α)
wr∗ = ,
(s − r + 1)(s − r + 2)
6(n − 1)(1 − α) − 2(s + 2r − 4)
ws∗ = , (11)
(s − r + 1)(s − r + 2)
s−j j−r
wj∗ = · wr + · ws , if j ∈ I{r+1,s−1} .
s−r s−r
(i) There exist λ∗1 , λ∗2 ∈ R and µ∗1 ≥ 0, . . . , µ∗n ≥ 0 such that,
n n−1
∂ 2 ∗ ∗
n−i
D (W ) + λ1 wi − 1 + λ2 · wi − α
∂wk n−1
i=1 i=1
n
∗
+ µj (−wj ) =0
j=1 W =W ∗
5
is positive definite on
X̂ = y h1 y T = 0, h2 y T = 0 and gj y T = 0 for all j with µj > 0 , (12)
where T
n−1 n−2 1
h1 = , , ..., ,0 , (13)
n−1 n−1 n−1
and
h2 = (1, 1, . . . , 1, 1)T . (14)
are the gradients of linear equality constraints, and
jth
gj = (0, 0, . . . , 0, −1 , 0, 0, . . . , 0)T (15)
is the gradient of the jth linear inequality constraint of problem (2).
6
We need to show that µ∗k ≥ 0 for k ∈
/ I{r,s} . That is,
(s − 1)(3k − 2s − 2)
α≥1− .
3(n − 1)(2k − s − 1)
and, therefore,
1 s−1
α≥1− ·
3 n−1
Finally, from the inequality
1 s−1 (s − 1)(3k − 2s − 2)
1− · ≥1−
3 n−1 3(n − 1)(2k − s − 1)
we get that (16) holds. The proof of the remaining case (r > 1 and s = n) is
carried out analogously.
(ii) The gradient vectors of linear equality and inequality constraints are com-
puted by (13), (14) and (15), respectively. If r = 1 and s = n then wj∗ = 0 for
all j = 1, . . . , n. Then it is easy to see that h1 and h2 are linearly independent. If
r = 1 and s < n then wj∗ = 0 for j = s + 1, . . . , n, and in this case
jth
gj = (0, 0, . . . , 0, −1 , 0, 0, . . . , 0)T ,
7
s + 2) of G is
n−s+1
1 0 ... 0
n−1
n−s n − s + 1
1 0 ... 0
n−1
1 1
n−s n − 1
(−1)n−s
n−s−1 = (−1) =
1 −1 . . . 0
n − s n − 1
n−1
1
.. .. .. . .
n − 1
. . . . 0
0 1 0 . . . −1
which means that the columns of G are linearly independent, and therefore, the
system
{h1 , h2 , gs+1 , . . . , gn },
is linearly independent.
If r > 1 and = n then wj∗ = 0 for j = 1, . . . , r − 1, and in this case
jth
gj = (0, 0, . . . , 0, −1 , 0, 0, . . . , 0)T ,
which means that the columns of F are linearly independent, and therefore, the
system
{h1 , h2 , g1 , . . . , gr−1 },
is linearly independent. So W ∗ is a regular point for problem (2).
(iii) Let us introduce the notation
n
n−1 n
2
n−i
K(W ) = D (W ) + λ∗1 wi − 1 + λ∗2 · wi − α + µ∗j (−wj ).
n−1
i=1 i=1 j=1
8
The Hessian matrix of K at W ∗ is
∂2 ∂2 2
2
K(W ) = D (W ) = · δkj ,
∂wk ∂wj W =W ∗ ∂wk ∂wj W =W ∗ n
where
1 if j = k
δjk =
0 otherwise.
That is,
1 0 ... 0
2 0 1 . . . 0
K ′′ (W ∗ ) = · . . . .
n .. .. . . ..
0 0 ... 1
which is a positive definite matrix on Rn .
n
n−i
X = W ∈ R W ≥ 0, wi = 1, · wi = α (17)
n−1
i=1 i=1
where X is the set of feasible solutions of problem (2). Taking into consideration
that D2 : Rn → R is a strictly convex, bounded and continuous function, and X
is a covex and compact subset of Rn , we can conclude that D2 attains its (unique)
global minimum on X at point W ∗ .
3 Example
In this Section we will determine the minimal variability five-dimensional weight-
ing vector under orness levels α = 0, 0.1, . . . , 0.9 and 1.0. First, we construct the
corresponding partition as
4
4
(0, 1) = Jr,5 ∪ J1,5 ∪ J1,s .
r=2 s=2
where
1 5−r−1 1 5−r 4−r 5−r
Jr,5 = · , · = , ,
3 5−1 3 5−1 12 12
for r = 2, 3, 4 and
1 5 − 2 1 10 − 1 3 9
J1,5 = · , · = , ,
3 5−1 3 5−1 12 12
and
1 s−1 1 s−2 13 − s 14 − s
J1,s = 1 − · , 1− · = , ,
3 5−1 3 5−1 12 12
9
for s = 2, 3, 4, and, therefore we get,
1 1 2 2 3 3 9
(0, 1) = 0, ∪ , ∪ , ∪ ,
12 12 12 12 12 12 12
9 10 10 11 11 12
∪ , ∪ , ∪ , .
12 12 12 12 12 12
Without loss of generality we can assume that α < 0.5, because if a weighting
vector W is optimal for problem (2) under some given degree of orness, α < 0.5,
then its reverse, denoted by W R , and defined as
wiR = wn−i+1
is also optimal for problem (2) under degree of orness (1 − α). Really, as was
shown by Yager [7], we find that
D2 (W R ) = D2 (W ) and orness(W R ) = 1 − orness(W ).
Therefore, for any α > 0.5, we can solve problem (2) by solving it for level of
orness (1 − α) and then taking the reverse of that solution.
Then we obtain the optimal weights from (11) as follows
• if α = 0.1 then
1 2
α ∈ J3,5 = , ,
12 12
and the associated minimal variablity weights are
w1∗ (0.1) = 0,
w2∗ (0.1) = 0,
2(10 + 3 − 2) − 6(5 − 1)(1 − 0.1) 0.4
w3∗ (0.1) = = = 0.0333,
(5 − 3 + 1)(5 − 3 + 2) 12
2
w5∗ (0.1) = − w3∗ (0.1) = 0.6334,
5−3+1
1 1
w4∗ (0.1) = · w3∗ (0.1) + · w5∗ (0.1) = 0.3333,
2 2
So,
W ∗ (α) = W ∗ (0.1) = (0, 0, 0.033, 0.333, 0.633)T ,
and, consequently,
W ∗ (0.9) = (W ∗ (0.1))R = (0.633, 0.333, 0.033, 0, 0)T .
with variance D2 (W ∗ (0.1)) = 0.0625.
10
• if α = 0.2 then
2 3
α ∈ J2,5 = ,
12 12
and in a similar manner we find that the associated minimal variablity weight-
ing vector is
and, therefore,
• if α = 0.3 then
3 9
α ∈ J1,5 = ,
12 12
and in a similar manner we find that the associated minimal variablity weight-
ing vector is
and, therefore,
• if α = 0.4 then
3 9
α ∈ J1,5 = ,
12 12
and in a similar manner we find that the associated minimal variablity weight-
ing vector is
and, therefore,
• if α = 0.5 then
11
4 Summary
We have extended the power of decision making with OWA operators by intro-
ducing considerations of minimizing variability into the process of selecting the
optimal alternative. Of particular significance in this work is the development of a
methodology to calculate the minimal variability weights in an analytic way.
References
[1] R.G. Brown, Smoothing, Forecasting and Prediction of Discrete Time Series
(Prentice-Hall, Englewood Cliffs, 1963).
[3] D. Filev and R. R. Yager, On the issue of obtaining OWA operator weights,
Fuzzy Sets and Systems, 94(1998) 157-169.
[7] R.R.Yager, Families of OWA operators, Fuzzy Sets and Systems, 59(1993)
125-148.
[9] R.R.Yager, On the analytic representation of the Leximin ordering and its
application to flexible constraint propagation, European Journal of Opera-
tional Research, 102(1997) 176-192.
12
Turku Centre for Computer Science
Lemminkäisenkatu 14
FIN-20520 Turku
Finland
http://www.tucs.abo.fi
University of Turku
• Department of Mathematical Sciences