You are on page 1of 16

On obtaining minimal variability

OWA operator weights

Robert Fullér
Department of Operations Research
Eötvös Loránd University
Pázmány Péter sétány 1C,
P. O. Box 120, H-1518 Budapest, Hungary
e-mail: rfuller@cs.elte.hu
and
Institute for Advanced Management Systems Research
Åbo Akademi University,
Lemminkäinengatan 14B, Åbo, Finland
e-mail: rfuller@mail.abo.fi
Péter Majlender
Turku Centre for Computer Sceince
Institute for Advanced Management Systems Research
Åbo Akademi University,
Lemminkäinengatan 14B, Åbo, Finland
e-mail: peter.majlender@mail.abo.fi

Turku Centre for Computer Science


TUCS Technical Report No 429
November 2001
ISBN 952-12-0911-9
TUCS ISSN 1239-1891
Abstract

One important issue in the theory of Ordered Weighted Averaging (OWA) oper-
ators is the determination of the associated weights. One of the first approaches,
suggested by O’Hagan, determines a special class of OWA operators having max-
imal entropy of the OWA weights for a given level of orness; algorithmically it is
based on the solution of a constrained optimization problem. Another considera-
tion that may be of interest to a decision maker involves the variability associated
with a weighting vector. In particular, a decision maker may desire low variability
associated with a chosen weighting vector. In this paper, using the Kuhn-Tucker
second-order sufficiency conditions for optimality, we shall analytically derive the
minimal variablity weighting vector for any level of orness.

Keywords: Multiple criteria analysis, Fuzzy sets, OWA operator, Lagrange multi-
plier
1 Introduction
The process of information aggregation appears in many applications related to
the development of intelligent systems. One sees aggregation in neural networks,
fuzzy logic controllers, vision systems, expert systems and multi-criteria decision
aids. In [6] Yager introduced a new aggregation technique based on the ordered
weighted averaging operators.
Definition 1.1. An OWA operator of dimension n is a mapping F : Rn → R that
has an associated weighting vector W = (w1 , . . . , wn )T of having the properties
w1 + · · · + wn = 1, 0 ≤ wi ≤ 1, i = 1, . . . , n, and such that
n

F (a1 , . . . , an ) = wi bi ,
i=1
where bj is the jth largest element of the collection of the aggregated objects
{a1 , . . . , an }.
A fundamental aspect of this operator is the re-ordering step, in particular an
aggregate ai is not associated with a particular weight wi but rather a weight is
associated with a particular ordered position of aggregate. When we view the OWA
weights as a column vector we shall find it convenient to refer to the weights with
the low indices as weights at the top and those with the higher indices with weights
at the bottom. It is noted that different OWA operators are distinguished by their
weighting function.
In [6], Yager introduced a measure of orness associated with the weighting
vector W of an OWA operator, defined as
n
 n−i
orness(W ) = · wi .
n−1
i=1
and it characterizes the degree to which the aggregation is like an or operation. It
is clear that orness(W ) ∈ [0, 1] holds for any weighting vector.
It is clear that the actual type of aggregation performed by an OWA operator
depends upon the form of the weighting vector [9]. A number of approaches have
been suggested for obtaining the associated weights, i.e., quantifier guided aggre-
gation [6, 7], exponential smoothing [3] and learning [10].
Another approach, suggested by O’Hagan [5], determines a special class of
OWA operators having maximal entropy of the OWA weights for a given level
of orness. This approach is based on the solution of he following mathematical
programming problem:
n
maximize disp(W ) = − wi ln wi
i=1
n
 n−i
subject to orness(W ) = · wi = α, 0 ≤ α ≤ 1 (1)
n−1
i=1
w1 + · · · + wn = 1, 0 ≤ wi , i = 1, . . . , n.

1
In [4] we transfered problem (1) to a polinomial equation which is then was
solved to determine the maximal entropy OWA weights.
Another interesting question is to determine the minimal variablity weighting
vector under given level of orness [8]. We shall meausure the variance of a given
weighting vector as
n
 1
D2 (W ) = · (wi − E(W ))2
n
i=1
n
1 n n
  2
1 2 1 1
= wi − wi = wi2 − 2 .
n n n n
i=1 i=1 i=1

where E(W ) = (w1 + · · · + wn )/ stands for the arithmetic mean of weights.


To obtain minimal variablity OWA weights under given level of orness we need
to solve the following constrained mathematical programming problem
n
1  1
minimize D2 (W ) = · wi2 − 2
n n
i=1
n
 n−i
subject to orness(w) = · wi = α, 0 ≤ α ≤ 1, (2)
n−1
i=1
w1 + · · · + wn = 1, 0 ≤ wi , i = 1, . . . , n.

Using the Kuhn-Tucker second-order sufficiency conditions for optimality, we shall


solve problem (2) analytically and derive the exact minimal variablity OWA weights
for any level of orness.

2 Obtaining minimal variability OWA weights


Let us consider the constrained optimization problem (2). First we note that if
n = 2 then from orness(w1 , w2 ) = α the optimal weights are uniquely defined as
w1∗ = α and w2∗ = 1 − α. Furthemore, if α = 0 or α = 1 then the associated
weighting vectors are uniquely defined as (0, 0, . . . , 0, 1)T and (1, 0, . . . , 0, 0)T ,
respectively, with variability

1 1
D2 (1, 0, . . . , 0, 0) = D2 (0, 0, . . . , 0, 1) = − 2.
n n

Suppose now that n ≥ 3 and 0 < α < 1. Let us

n n
  n
 
1 1 n−i
L(W, λ1 , λ2 ) = wi2 − 2 + λ1 wi − α + λ2 wi − 1 .
n n n−1
i=1 i=1 i=1

2
denote the Lagrange function of constrained optimization problem (2), where
λ1 and λ2 are real numbers. Then the partial derivatives of L are computed as

∂L 2wj n−j
= + λ1 + · λ2 = 0, 1 ≤ j ≤ n, (3)
∂wj n n−1
n
∂L 
= wi − 1 = 0,
∂λ1
i=1
n
∂L  n−i
= · wi − α = 0.
∂λ2 n−1
i=1

We shall suppose that the optimal weighting vector has the following form

W = (0, . . . , 0, wp , . . . , wq , 0 . . . , 0)T (4)

where 1 ≤ p < q ≤ n and use the notation

I{p,q} = {p, p + 1, . . . , q − 1, q},

for the indexes from p to q. So, wj = 0 if j ∈


/ I{p,q} and wj ≥ 0 if j ∈ I{p,q} .
For j = p we find that

∂L 2wp n−p
= + λ1 + · λ2 = 0
∂wp n n−1
and for j = q we get

∂L 2wq n−q
= + λ1 + · λ2 = 0
∂wq n n−1
That is,
2(wp − wq ) q − p
+ · λ2 = 0
n n−1
and therefore, the optimal values of λ1 and λ2 (denoted by λ∗1 and λ∗2 ) should
satisfy the following equations
 
∗ 2 n−q n−p n−1 2
λ1 = · wp − · wq and λ∗2 = · · (wq − wp ). (5)
n q−p q−p q−p n
Substituting λ∗1 for λ1 and λ∗2 for λ2 in (3) we get
 
2 2 n−q n−p n−j n−1 2
· wj + · wp − · wq + · · · (wq − wp ) = 0.
n n q−p q−p n−1 q−p n
That is the jth optimal weight should satisfy the equation

q−j j−p
wj∗ = · wp + · wq , j ∈ I{p,q} . (6)
q−p q−p

3
From representation (4) we get
q

wi∗ = 1,
i=p

That is,
q  
 q−i i−p
· wp + · wq = 1,
q−p q−p
i=p
i.e.
2
wp + wq =.
q−p+1
From the constraint orness(w) = α we find
q q q
 n−i  n−i q−i  n−i i−p
· wi = · · wp + · · wq = α
n−1 n−1 q−p n−1 q−p
i=p i=p i=p

that is,
2(2q + p − 2) − 6(n − 1)(1 − α)
wp∗ = . (7)
(q − p + 1)(q − p + 2)
and
2 6(n − 1)(1 − α) − 2(q + 2p − 4)
wq∗ = − wp∗ = . (8)
q−p+1 (q − p + 1)(q − p + 2)
The optimal weighting vector
W ∗ = (0, . . . , 0, wp∗ , . . . , wq∗ , 0 . . . , 0)T
is feasible if and only if wp∗ , wq∗ ∈ [0, 1], because according to (6) any other wj∗ ,
j ∈ I{p,q} is computed as their convex linear combination.
Using formulas (7) and (8) we find
 
1 2q + p − 2 1 q + 2p − 4
wp∗ , wq∗ ∈ [0, 1] ⇐⇒ α ∈ 1 − · ,1 − ·
3 n−1 3 n−1
The following (disjunctive) partition of the unit interval (0, 1) will be crucial in
finding an optimal solution to problem (2):
n−1
 n−1

(0, 1) = Jr,n ∪ J1,n ∪ J1,s . (9)
r=2 s=2
where
 
1 2n + r − 2 1 2n + r − 3
Jr,n = 1− · , 1− · , r = 2, . . . , n − 1,
3 n−1 3 n−1
 
1 2n − 1 1 n−2
J1,n = 1− · , 1− · ,
3 n−1 3 n−1
 
1 s−1 1 s−2
J1,s = 1− · , 1− · , s = 2, . . . , n − 1.
3 n−1 3 n−1

4
Consider again problem (2) and suppose that α ∈ Jr,s for some r and s from
partition (9). Such r and s always exist for any α ∈ (0, 1), furthermore, r = 1 or
s = n should hold.
Then
W ∗ = (0, . . . , 0, wr∗ , . . . , ws∗ , 0 . . . , 0)T (10)
where

wj∗ = 0, if j ∈
/ I{r,s} ,
2(2s + r − 2) − 6(n − 1)(1 − α)
wr∗ = ,
(s − r + 1)(s − r + 2)
6(n − 1)(1 − α) − 2(s + 2r − 4)
ws∗ = , (11)
(s − r + 1)(s − r + 2)
s−j j−r
wj∗ = · wr + · ws , if j ∈ I{r+1,s−1} .
s−r s−r

and I{r+1,s−1} = {r + 1, . . . , s − 1}. Furthermore, from the construction of W ∗ it


is clear that
n  s

wi = wi∗ = 1, wi∗ ≥ 0, i = 1, 2, . . . , n,
i=1 i=r

and orness(W ∗ )= α, that is, W ∗ is feasible for problem (2).


We will show now that W ∗ , defined by (10), satisfies the Kuhn-Tucker second-
order sufficiency conditions for optimality ([2], page 58). Namely,

(i) There exist λ∗1 , λ∗2 ∈ R and µ∗1 ≥ 0, . . . , µ∗n ≥ 0 such that,
 n   n−1 
∂ 2 ∗ ∗
 n−i
D (W ) + λ1 wi − 1 + λ2 · wi − α
∂wk n−1
i=1 i=1
n
 


+ µj (−wj )  =0
j=1 W =W ∗

for 1 ≤ k ≤ n and µ∗j wj∗ = 0, j = 1, . . . , n.

(ii) W ∗ is a regular point of the constraints,

(iii) The Hessian matrix,


 n   n−1 
∂2 2 ∗ ∗
 n−i
D (W ) + λ1 wi − 1 + λ2 · wi − α
∂W 2 n−1
i=1 i=1
 n 


+ µj (−wj ) 
j=1 W =W ∗

5
is positive definite on



X̂ = y h1 y T = 0, h2 y T = 0 and gj y T = 0 for all j with µj > 0 , (12)

where  T
n−1 n−2 1
h1 = , , ..., ,0 , (13)
n−1 n−1 n−1
and
h2 = (1, 1, . . . , 1, 1)T . (14)
are the gradients of linear equality constraints, and
jth

gj = (0, 0, . . . , 0, −1 , 0, 0, . . . , 0)T (15)
is the gradient of the jth linear inequality constraint of problem (2).

Proof. (i) According to (5) we get


 
∗ 2 n−s ∗ n−r ∗ n−1 2
λ1 = · · wr − · ws and λ∗2 = · · (ws∗ − wr∗ )
n s−r s−r s−r n
and
2 n−k ∗
· wk∗ + λ∗1 + · λ − µk = 0.
n n−1 2
for k = 1, . . . , n. If k ∈ I{r,s} then
   
∗ 2 s−k ∗ k−r ∗ 2 n−s ∗ n−r ∗
µk = · · wr + · ws + · · wr − · ws
n s−r s−r n s−r s−r
n−k n−1 2
+ · · · (ws∗ − wr∗ )
n−1 s−r n
2 1  
= · (s − k + n − s − n + k)wr∗ + (k − r − n + r + n − k)ws∗
n s−r
= 0.
/ I{r,s} then wk∗ = 0. Then from the equality
If k ∈
n−k ∗
λ∗1 + · λ − µk = 0,
n−1 2
we find
n−k ∗
µ∗k = λ∗1 + ·λ
n−1 2
 
2 n−s ∗ n−r ∗ n−k n−1 2
= · · wr − · ws + · · · (ws∗ − wr∗ )
n s−r s−r n−1 s−r n
2 1  
= · · (k − s)wr∗ + (r − k)ws∗ .
n s−r

6
We need to show that µ∗k ≥ 0 for k ∈
/ I{r,s} . That is,

2(2s + r − 2) − 6(n − 1)(1 − α)


(k − s)wr∗ + (r − k)ws∗ = (k − s) ·
(s − r + 1)(s − r + 2)
(16)
6(n − 1)(1 − α) − 2(s + 2r − 4)
+(r − k) · ≥ 0.
(s − r + 1)(s − r + 2)

If r = 1 and s = n then we get that µ∗k = 0 for k = 1, . . . , n. Suppose now that


r = 1 and s < n. In this case the inequality k > s > 1 should hold and (16) leads
to the following requirement for α,

(s − 1)(3k − 2s − 2)
α≥1− .
3(n − 1)(2k − s − 1)

On the other hand, from α ∈ J1,s and s < n we have


 
1 s−1 1 s−2
α∈ 1− · , 1− · ,
3 n−1 3 n−1

and, therefore,
1 s−1
α≥1− ·
3 n−1
Finally, from the inequality

1 s−1 (s − 1)(3k − 2s − 2)
1− · ≥1−
3 n−1 3(n − 1)(2k − s − 1)

we get that (16) holds. The proof of the remaining case (r > 1 and s = n) is
carried out analogously.
(ii) The gradient vectors of linear equality and inequality constraints are com-
puted by (13), (14) and (15), respectively. If r = 1 and s = n then wj∗ = 0 for
all j = 1, . . . , n. Then it is easy to see that h1 and h2 are linearly independent. If
r = 1 and s < n then wj∗ = 0 for j = s + 1, . . . , n, and in this case

jth

gj = (0, 0, . . . , 0, −1 , 0, 0, . . . , 0)T ,

for j = s + 1, . . . , n. Consider the matrix

G = [hT1 , hT2 , gs+1


T
, . . . , gnT ] ∈ Rn×(n−s+2) .

Then the determinant of the lower-left submatrix of dimension (n − s + 2) × (n −

7
s + 2) of G is
 
 n−s+1 
 1 0 ... 0 
 
 n−1   
 n−s  n − s + 1 
 1 0 ... 0  

 n−1
  1 1
n−s  n − 1
(−1)n−s

 n−s−1  = (−1) =
 1 −1 . . . 0  
 n − s  n − 1

 n−1
  1
 .. .. .. . .

 n − 1
 . . . . 0 
 
 0 1 0 . . . −1 

which means that the columns of G are linearly independent, and therefore, the
system
{h1 , h2 , gs+1 , . . . , gn },
is linearly independent.
If r > 1 and = n then wj∗ = 0 for j = 1, . . . , r − 1, and in this case

jth

gj = (0, 0, . . . , 0, −1 , 0, 0, . . . , 0)T ,

for j = 1, . . . , r − 1. Consider the matrix

F = [hT1 , hT2 , g1T , . . . , gr−1


T
] ∈ Rn×(r+1) .

Then the determinant of the upper-left submatrix of dimension (r + 1) × (r + 1)


of F is
 
 n−1 
 1 −1 . . . 0 

 n−1 
 .. .. .. . .   

 . . . . 0 
  n−r 
 n−r+1  1 
   1

 n−1 1 0 . . . −1  = (−1)
 r−1  n−1 = (−1)r−1
  
 n − r − 1  n − 1
 n−r   1

 n−1 1 0 ... 0  n − 1
 
 n−r−1 

 n−1 1 0 . . . 0 


which means that the columns of F are linearly independent, and therefore, the
system
{h1 , h2 , g1 , . . . , gr−1 },
is linearly independent. So W ∗ is a regular point for problem (2).
(iii) Let us introduce the notation
n
   n−1  n
2
 n−i
K(W ) = D (W ) + λ∗1 wi − 1 + λ∗2 · wi − α + µ∗j (−wj ).
n−1
i=1 i=1 j=1

8
The Hessian matrix of K at W ∗ is
 
∂2  ∂2 2
 2
K(W ) = D (W ) = · δkj ,
∂wk ∂wj W =W ∗ ∂wk ∂wj W =W ∗ n
where
1 if j = k
δjk =
0 otherwise.
That is,  
1 0 ... 0
2 0 1 . . . 0

K ′′ (W ∗ ) = ·  . . . . 
n  .. .. . . .. 
0 0 ... 1
which is a positive definite matrix on Rn .

So, the objective function D2 (W ) has a local minimum at point W = W ∗ on


 n n


n
  n−i
X = W ∈ R W ≥ 0, wi = 1, · wi = α (17)
n−1
i=1 i=1

where X is the set of feasible solutions of problem (2). Taking into consideration
that D2 : Rn → R is a strictly convex, bounded and continuous function, and X
is a covex and compact subset of Rn , we can conclude that D2 attains its (unique)
global minimum on X at point W ∗ .

3 Example
In this Section we will determine the minimal variability five-dimensional weight-
ing vector under orness levels α = 0, 0.1, . . . , 0.9 and 1.0. First, we construct the
corresponding partition as
4
 4

(0, 1) = Jr,5 ∪ J1,5 ∪ J1,s .
r=2 s=2

where    
1 5−r−1 1 5−r 4−r 5−r
Jr,5 = · , · = , ,
3 5−1 3 5−1 12 12
for r = 2, 3, 4 and
   
1 5 − 2 1 10 − 1 3 9
J1,5 = · , · = , ,
3 5−1 3 5−1 12 12
and    
1 s−1 1 s−2 13 − s 14 − s
J1,s = 1 − · , 1− · = , ,
3 5−1 3 5−1 12 12

9
for s = 2, 3, 4, and, therefore we get,
       
1 1 2 2 3 3 9
(0, 1) = 0, ∪ , ∪ , ∪ ,
12 12 12 12 12 12 12
     
9 10 10 11 11 12
∪ , ∪ , ∪ , .
12 12 12 12 12 12
Without loss of generality we can assume that α < 0.5, because if a weighting
vector W is optimal for problem (2) under some given degree of orness, α < 0.5,
then its reverse, denoted by W R , and defined as
wiR = wn−i+1
is also optimal for problem (2) under degree of orness (1 − α). Really, as was
shown by Yager [7], we find that
D2 (W R ) = D2 (W ) and orness(W R ) = 1 − orness(W ).
Therefore, for any α > 0.5, we can solve problem (2) by solving it for level of
orness (1 − α) and then taking the reverse of that solution.
Then we obtain the optimal weights from (11) as follows

• if α = 0 then W ∗ (α) = W ∗ (0) = (0, 0, . . . , 0, 1)T and, therefore,


W ∗ (1) = (W ∗ (0))R = (1, 0, . . . , 0, 0)T .

• if α = 0.1 then  
1 2
α ∈ J3,5 = , ,
12 12
and the associated minimal variablity weights are
w1∗ (0.1) = 0,
w2∗ (0.1) = 0,
2(10 + 3 − 2) − 6(5 − 1)(1 − 0.1) 0.4
w3∗ (0.1) = = = 0.0333,
(5 − 3 + 1)(5 − 3 + 2) 12
2
w5∗ (0.1) = − w3∗ (0.1) = 0.6334,
5−3+1
1 1
w4∗ (0.1) = · w3∗ (0.1) + · w5∗ (0.1) = 0.3333,
2 2
So,
W ∗ (α) = W ∗ (0.1) = (0, 0, 0.033, 0.333, 0.633)T ,
and, consequently,
W ∗ (0.9) = (W ∗ (0.1))R = (0.633, 0.333, 0.033, 0, 0)T .
with variance D2 (W ∗ (0.1)) = 0.0625.

10
• if α = 0.2 then  
2 3
α ∈ J2,5 = ,
12 12
and in a similar manner we find that the associated minimal variablity weight-
ing vector is

W ∗ (0.2) = (0.0, 0.04, 0.18, 0.32, 0.46)T ,

and, therefore,

W ∗ (0.8) = (0.46, 0.32, 0.18, 0.04, 0.0)T ,

with variance D2 (W ∗ (0.2)) = 0.0296.

• if α = 0.3 then  
3 9
α ∈ J1,5 = ,
12 12
and in a similar manner we find that the associated minimal variablity weight-
ing vector is

W ∗ (0.3) = (0.04, 0.12, 0.20, 0.28, 0.36)T ,

and, therefore,

W ∗ (0.7) = (0.36, 0.28, 0.20, 0.12, 0.04)T ,

with variance D2 (W ∗ (0.3)) = 0.0128.

• if α = 0.4 then  
3 9
α ∈ J1,5 = ,
12 12
and in a similar manner we find that the associated minimal variablity weight-
ing vector is

W ∗ (0.4) = (0.12, 0.16, 0.20, 0.24, 0.28)T ,

and, therefore,

W ∗ (0.6) = (0.28, 0.24, 0.20, 0.16, 0.12)T ,

with variance D2 (W ∗ (0.4)) = 0.0032.

• if α = 0.5 then

W ∗ (0.5) = (0.2, 0.2, 0.2, 0.2, 0.2)T .

with variance D2 (W ∗ (0.5)) = 0.

11
4 Summary
We have extended the power of decision making with OWA operators by intro-
ducing considerations of minimizing variability into the process of selecting the
optimal alternative. Of particular significance in this work is the development of a
methodology to calculate the minimal variability weights in an analytic way.

References
[1] R.G. Brown, Smoothing, Forecasting and Prediction of Discrete Time Series
(Prentice-Hall, Englewood Cliffs, 1963).

[2] V. Chankong and Y. Y. Haimes, Multiobjective Decision Making: Theory


and Methodology (North-Holland, 1983).

[3] D. Filev and R. R. Yager, On the issue of obtaining OWA operator weights,
Fuzzy Sets and Systems, 94(1998) 157-169.

[4] R. Fullér and P. Majlender, An analytic approach for obtaining maximal


entropy OWA operator weights, Fuzzy Sets and Systems, 124(2001) 53-57.

[5] M. O’Hagan, Aggregating template or rule antecedents in real-time expert


systems with fuzzy set logic, in: Proc. 22nd Annual IEEE Asilomar Conf.
Signals, Systems, Computers, Pacific Grove, CA, 1988 81-689.

[6] R.R.Yager, Ordered weighted averaging aggregation operators in multi-


criteria decision making, IEEE Trans. on Systems, Man and Cybernetics,
18(1988) 183-190.

[7] R.R.Yager, Families of OWA operators, Fuzzy Sets and Systems, 59(1993)
125-148.

[8] R.R.Yager, On the Inclusion of variance in decision making under un-


certainty, International Journal of Uncertainty, Fuzziness and Knowledge-
Based Systems, 4(1995) 401-419.

[9] R.R.Yager, On the analytic representation of the Leximin ordering and its
application to flexible constraint propagation, European Journal of Opera-
tional Research, 102(1997) 176-192.

[10] R. R. Yager and D. Filev, Induced ordered weighted averaging operators,


IEEE Trans. on Systems, Man and Cybernetics – Part B: Cybernetics,
29(1999) 141-150.

12
Turku Centre for Computer Science
Lemminkäisenkatu 14
FIN-20520 Turku
Finland
http://www.tucs.abo.fi

University of Turku
• Department of Mathematical Sciences

Åbo Akademi University


• Department of Computer Science
• Institute for Advanced Management Systems Research

Turku School of Economics and Business Administration


• Institute of Information Systems Science

You might also like