You are on page 1of 11

Articles

Multiobjective
Optimization

Matthias Ehrgott

A
n Using some real-world examples I
illustrate the important role of multiob-
jective optimization in decision making n investor composes a portfolio of stocks in order to obtain a high
and its interface with preference han- return on his or her investment with a small risk of incurring a loss; an
dling. I explain what optimization in oncologist prescribes radiotherapy to a cancer patient so as to destroy the
the presence of multiple objectives
tumor without causing damage to healthy organs; an airline manager
means and discuss some of the most
common methods of solving multiob- constructs schedules that incur small salary costs and that ensure smooth
jective optimization problems using operation even in the case of disruptions. All three decision makers
transformations to single-objective (DMs) are in a similar situation—they need to make a decision trying to
optimization problems. Finally, I achieve several conflicting goals at the same time: The highest return
address linear and combinatorial opti- investments are in general the riskiest ones, tumors can always be
mization problems with multiple objec- destroyed at the expense of irreversible damage to healthy organs, and
tives and summarize techniques for
the cheapest schedules to operate are ones that leave as little as possible
solving them. Throughout the article I
refer to the real-world examples intro-
time between flights, wreaking havoc to operations in the case of unex-
duced at the beginning. pected delays.
Moreover, the investor, the oncologist, and the airline manager are all
in a situation where the number of available options or alternatives is
very large or even infinite. There are infinitely many ways to invest mon-
ey and infinitely many possible radiotherapy treatments, but the num-
ber of feasible crew schedules is finite, albeit astronomical in practice.
The alternatives are therefore described by constraints, rather than
explicitly known: the sums invested in every stock must equal the total
invested; the radiotherapy treatment must meet physical and clinical
constraints; crew schedules must ensure that each flight has exactly one
crew assigned to operate it.
Mathematically, the alternatives are described by vectors in variable
or decision space; the set of all vectors satisfying the constraints is called
the feasible set in decision space. The consequences or attributes of the
alternatives are described as vectors in objective or outcome space, where
outcome (objective) vectors are a function of the decision (variable) vec-
tors. The set of outcomes corresponding to feasible alternatives is called

Copyright © 2008, Association for the Advancement of Artificial Intelligence. All rights reserved. ISSN 0738-4602 WINTER 2008 47
Articles

the feasible set in objective space. The decision proofs of all statements can be found in my text-
problem consists in finding that alternative with book (Ehrgott 2005b).
the most preferred outcome. But what exactly does
“most preferred outcome” mean? Although in
each of the attributes (or objectives or goals or cri-
Multiobjective
teria) the answer is clear (high return is preferred to Optimization Problems
low, cheap schedules are preferred to expensive
Following the description above, I will assume that
ones), the situation is more difficult when all crite-
alternatives can be described by vectors x 僆 n, the
ria are considered together: It is not possible to
decision space. Feasible alternatives are those that
compare investments if the first has higher return
satisfy certain constraints, mathematically
but also higher risk than the second unless further
expressed as g(x)  0, where g : n → m is a func-
information on trade-offs between the objectives tion describing the m constraints. I will refer to X
or other preference information is available. One as feasible set in decision space and to elements of X
can distinguish three situations. as feasible solutions. The outcome vector y = f(x)
If the decision maker is able to completely spec- associated with alternative x is an evaluation of x
ify his or her preferences explicitly, it is possible to according to a function f : n → p, that is, fk(x) is
construct a utility function that combines the cri- the value of objective k attained by solution x and
teria in a single function using multiattribute util- there are p objectives in total. The set of outcome
ity theory (see chapter seven in Figueira, Greco, vectors of all feasible solutions, Y = {f(x): x 僆 X}, is
and Ehrgott [2005]). The decision problem then called the feasible set in objective space p of the
turns into a single-objective optimization problem MOP (1).
that can then be solved by traditional mathemati- Because smaller values are preferred for all objec-
cal programming methods. This scenario is very tives, a multiobjective optimization problem con-
unrealistic. sists in the minimization of f over all feasible solu-
If preference information is not complete or not tions x in the feasible set X = {x 僆 n : g(x)  0}:
explicitly available but one assumes that the DM is
min{f(x): g(x)  0, x 僆 n}. (1)
implicitly aware of those preferences, one can
involve the DM in the solution process and assess From now on I will refer to the constraints just
preferences by asking for pairwise comparisons, as x 僆 X. To solve problem 1 it is not enough to
aspiration and reservation levels, and so on. Such consider minima of the individual objective func-
a scenario leads to interactive methods for finding tions fk, that is x* 僆 X such that fk (x*)  fk (x) for
a preferred alternative, where preference elicitation all x 僆 X.
from the DM alternates with some calculation, In multiobjective optimization, illustrations in
objective space are very enlightening. Figure 1
often the optimization of a function using the
shows the feasible set Y in objective space of a
information given by the DM as parameters (see
problem with two objectives. The minimum value
chapter 16 in Figueira, Greco, and Ehrgott [2005]).
of f1 is 0, attained for a solution x with f2(x) = 8.
If, however, no preference information is avail-
The minimum value of f2 is 1, and there are sever-
able, DMs face a multiobjective optimization problem
al minimizers with values for f1 between seven and
(chapter 17 in Figueira, Greco, and Ehrgott [2005]).
eight. As in figure 1, the set of minimizers of all p
This is what I am interested in this article. The only
functions is usually disjoint.
assumption is that for each of the criteria the DM
In fact, because problem 1 implies minimization
prefers less to more (which is no loss of generality
over vectors, it is necessary to specify a (partial)
because in the case of more is better one can switch
order on p to define the meaning of the min oper-
the sign of the objective). The goal of multiobjec-
ator (unlike the set of real numbers , p does not
tive optimization is to identify those alternatives
have a canonical order). I use the following nota-
that cannot be improved according to the “less is
tions:
better” paradigm. These alternatives are then sub-
y1  y2 ⇔ y1k  y2k for k = 1, …, p;
ject to further analysis to find the most preferred
solution for the DM. This latter step is the topic of y1 < y2 ⇔ y1k < y2k for k = 1, …, p;
multicriteria decision analysis (Figueira, Greco,
y1 ≤ y2 ⇔ y1  y2 and y1 ≠ y2;
and Ehrgott 2005).
In the following section I define multiobjective y1 lex y2 ⇔ y1 = y2 or y11 < y21 for the smallest index
optimization a little more formally and explain l with y1l ≠ y2l}.
what solving such a problem means. Then I talk Thus, outcome vectors are compared compo-
about some solution techniques before focusing on nentwise. Weak dominance  requires all compo-
multiobjective optimization problems with linear nents of y1 to be less than or equal to those of y2.
objectives and constraints and either continuous Strict dominance < asks for strict inequality in all
or discrete variables in some more detail. A math- components and dominance ≤ for strict inequality
ematical treatment of what follows including in at least one component. Finally, for the lexico-

48 AI MAGAZINE
Articles

graphic order, the first component for which y1


and y2 differ is decisive.
Using the lexicographic order lex, x* is lexico-
graphically minimal if f(x*) lex f(x) for all x 僆 X. 10
This definition relates to the given order of objec-
tives and implies that the objectives are ordered 9
according to decreasing importance. A more gen- 8
eral definition of lexicographic minimality is to
define x* to be lexicographically minimal if fπ(x*) 7
lex fπ(x) for all x 僆 X and some permutation fπ of
6
(f1, …, fp). The lexicographic minima in figure 1 are Y = f(X)
(0, 8) and (7, 1). 5
The lexicographic order is total, that is, any two
vectors can be compared. Multiobjective optimiza- 4
tion most often uses the partial orders ≤ and < , for 3
which this is not the case (hence the name partial
order). A feasible solution x* is weakly efficient if 2
there is no x 僆 X with f(x) < f(x*). f(x*) is then
1
called weakly nondominated point. x* belongs to the
set of efficient solutions XE if there is no x with f(x) 0
≤ f(x*). In that case f(x*) 僆 YN = f(XE) is nondomi- 0 1 2 3 4 5 6 7 8 9 10
nated. In figure 1 the point (2, 7) is (strictly) dom-
inated, as there are feasible points to the left and
down. Point (4, 4), however, is nondominated
whereas (4, 5) is weakly nondominated but domi- Figure 1. Feasible Set in Objective Space of a Multiobjective
nated by (4, 4). Optimization Problem with Two Objectives.
In other words, x* 僆 X is efficient if a move to
another solution x that improves one objective
implies that at least one other deteriorates. Thus,
efficient solutions and nondominated points are
about trade-offs between the different objectives.
In order to avoid unbounded trade-offs, which are points even though every feasible point is non-
undesirable in most practical applications, the fol- dominated.
lowing definition is often used. Having defined the meaning of min in problem
A feasible solution x* is properly efficient if x* is effi- 1 and thus the concept of optimality in multiob-
cient and there is a scalar M > 0 such that for each k jective optimization, I briefly mention conditions
and x with fk(x) < fk(x*) there is l with fl(x*) < fl(x) for the existence of efficient solutions respectively
and (fk(x*) – fk(x)) / (fl(x) – fl(x*))  M. The outcome nondominated points. There are a variety of con-
f(x*) is called properly nondominated. ditions known in the literature. Just as in the sin-
Figure 2 shows weakly nondominated points, gle-objective case they usually use some form of
nondominated points, and properly nondominat- compactness of X and continuity of f, respectively
ed points for the same example used in figure 1. compactness of Y. Borwein (1983) proved that a
The points on the curves between (0, 8) and (1.2, nondominated point exists if there is some feasi-
6) as well as (3.6, 6) and (8, 1) are all weakly non- ble point y0 such that that the set of feasible points
dominated. Among those, only points between (0, weakly dominating y0 is compact. In figure 2 this is
8) and (1.2, 6), between (3.6, 6) and (4, 5) (exclud- shown for y0 = (2, 7) and y0 = (4, 4). The result
ing those points), and between (4, 4) and (7, 1) are implies that an efficient solution to problem 1
nondominated. Among the nondominated points, exists if X is compact and the function f satisfies a
(1.2, 6) is not properly nondominated. weak continuity condition.
From the definitions it is clear that each proper- The next question is the range of values of the
ly efficient solution is efficient and that each effi- objective functions fk over the efficient set. The ide-
cient solution is weakly efficient. Correspondingly, al point is a vector of the best (minimal) values
each properly nondominated point is nondomi- that each fk can take over the feasible, and thus the
nated and each nondominated point is weakly efficient set: ykI = min{fk(x):x 僆 X}. The nadir point,
nondominated. Figure 2 shows that the inclusions on the other hand, gives the worst outcomes over
are strict in general. In fact, it is possible that there the efficient set, yNK = max{fk(x):x 僆 XE}. The maxi-
is only one nondominated point among an infi- mum of fk over efficient solutions is different from
nite set of weakly nondominated ones. Also, it is the maximum over feasible solutions, the latter
possible that there are no properly nondominated defining the antideal point by yAIk = max{fk(x):x 僆

WINTER 2008 49
Articles

gle-objective optimization tools. The principle of


scalarization is to convert the multiobjective pro-
gram (1) to a single-objective program that usually
10 depends on some parameters not included in (1)
yAI and then solve the scalarized problem repeatedly
9
8
yN with different parameter values. A scalarization
method ideally has the properties of correctness,
that is, an optimal solution to the scalarized prob-
7 lem is a (weakly, properly) efficient solution to (1)
6 and completeness, that is, every (weakly, properly)
Y = f(X) efficient solution to (1) can be found by solving a
5 scalarized problem with appropriate parameter val-
ues.
4
There are scores of scalarization techniques in
3 the literature. Most of them follow one ore more of
three ideas: minimize an aggregation of the objec-
2 tives, convert objectives to constraints, or mini-
yI mize the distance to some reference point. In the
1
rest of this section I explain these three ideas with
0 the best known scalarization techniques.
0 1 2 3 4 5 6 7 8 9 10
The most straightforward idea to scalarize an
MOP is to assign some nonnegative weight (inter-
preted as importance of the objective) to each
objective and add the weighted objectives up to be
minimized. This is known as the weighted-sum
Figure 2. Weakly Nondominated, Nondominated,
method. With weight vector ␭ ≥ 0 the single-objec-
and Properly Nondominated Points.
tive problem to minimize the weighted sum of the
objects subject to the original feasibility con-
straints is

⎪⎧ p ⎪⎫
X}. Apart from yI, yN, and yAI, utopian points yU = yI min ⎪⎨∑ λ k f k ( x) : x ∈ X ⎪⎬. (2)
– ⑀ for some vector ⑀ 僆 p with small positive ⎪⎪⎩ k=1 ⎪⎪⎭
entries are often used. Clearly, yI  y  yN for all y
The (negative of) weight vector ␭ is the normal
僆 YN and under the assumptions of conflicting
of a hyperplane in objective space, and can be
objectives, continuity and compactness, for each k
interpreted as a direction of minimization as
= 1, …, p there is y 僆 YN such that yk = ykI (yk = yNk).
shown in figure 3 for ␭ = (1, 1), ␭ = (1, 0), and ␭ =
Thus, the ideal and nadir points are the best possi-
(0, 1).
ble lower and upper bound vectors for the non-
Results concerning the correctness of the
dominated set. Figure 2 illustrates these points for
method depend on the values of ␭. Assuming x* to
my little example. Any point strictly to the left and
be an optimal solution to problem 2 it is easy to
below yI = (0, 1) is a utopian point.
see that the following assertions hold.
For the rest of this article I will assume that XE
(and thus YN) is nonempty and that yI ≠ yN, that is, If ␭ ≥ 0 then x* is a weakly efficient solution to prob-
lem 1.
problem 1 is not trivial. The multiobjective opti-
mization problems faced by the investor, the If furthermore there is no other feasible point y
oncologist, and the airline manager are then to such that the sum of ␭k yk equals the sum of ␭k fk(x*)
identify portfolios for which an increase in return then x* is an efficient solution to problem 1.
can only be achieved by accepting higher risk; find If all weights are positive then x* is a properly effi-
treatment plans that cannot provide better protec- cient solution to 1.
tion of healthy organs without compromising
It is important to note that the weighted-sum
tumor control; and determine crew schedules for
method is not complete in general. There is no
which reducing cost implies greater delays during
nonnegative nonzero weight vector that can estab-
disruptions.
lish (4, 4) as a nondominated point of the small
example in figure 3 by solving a weighted-sum
Finding Efficient Solutions problem. For a completeness result, convexity
assumptions are needed. Geoffrion (1968) proved
by Scalarization that if X and f are such that that the set Y, extend-
In this section I will talk about methods to solve ed by all points in p that are weakly dominated
multiobjective optimization problems using sin- by at least one point in Y, is convex then for any

50 AI MAGAZINE
Articles

weakly efficient solution x* there is a weight vector


␭ ≥ 0 such that x* is an optimal solution to prob-
lem 2; and for any properly efficient solution x*
there is a positive ␭ > 0 such that x* is an optimal 10
solution to problem 2.
9
Note that there is no sufficient condition for effi-
cient solutions. Although positive weights will 8
always yield properly efficient solutions, some zero λ = (1, 0)
weights may be needed to obtain efficient solu- 7
tions with unbounded trade-offs. Because weight 6
vectors ␭ ≥ 0 may, however, result in weakly non-
5
Y = f(X)
dominated points (such as for ␭ = (0, 1) in figure 3)
it is not possible to give a characterization of effi- 4
cient solutions. In fact, in multiobjective opti-
mization it is common that statements about 3
weakly and properly efficient solutions can be
2
proved, whereas analogous results for efficient λ = (1, 1)
solutions that are “in between” weakly and prop- 1
erly efficient ones are missing. λ = (0, 1)
To address the issue that some efficient solutions 0
0 1 2 3 4 5 6 7 8 9 10
cannot be found by solving a weighted-sum prob-
lem, the notion of supported efficient solutions,
the set of feasible solutions that are optimal to a
weighted-sum problem with positive weighting
vector, is useful. Geometrically, supported efficient Figure 3. Solving a Weighted-Sum Problem.
solutions are efficient solutions with f(x) on the
“lower left” boundary of the convex hull of Y as
shown in figure 3.
I have mentioned earlier that the difference
between nondominated and properly nondomi-
nated points can be large. For convex problems, to analyze the efficient portfolios to make a final
however, Hartley (1978) has shown that difference decision, taking into account personal, that is sub-
can only occur on the boundary of the nondomi- jective, trade-offs between risk and return.
nated set: If Y extended by all points dominated by To solve nonconvex MOPs another idea is need-
some y 僆 Y is closed and convex then the set of ed. The most popular method is to retain only one
properly nondominated points, which is equal to of the p objectives and turn all others into con-
the set of outcomes of supported efficient solu- straints, a concept known as ⑀-constraint method.
tions, is contained in the set of nondominated For some vector ⑀ 僆 p of upper bounds the scalar
points, which in turn is contained in the closure of optimization problem becomes
the set of outcomes of the set of supported efficient min{fl(x): fk(x)  ⑀k for k ≠ l, x 僆 X}. (3)
solutions. It is easy to see that the ⑀-constraint method is
The statements mentioned above imply that the
correct. It is even complete and does not require
weighted-sum method is unsuitable for solving
any convexity assumption. Chankong and Haimes
nonconvex MOPs. However, because the ␭’s are
(1983) proved the following statements:
often interpreted as importance weights (that is,
If x* is an optimal solution to problem 3 then x* is
preference information), the temptation to use it
weakly efficient.
in practice is strong. The lack of mathematical jus-
tification should be a warning not to succumb to If x* is an optimal solution to problem 3 and all
this temptation and to handle preference informa- optimal solutions have the same objective value
vector y = f(x*) then x* is efficient.
tion with care in multiobjective optimization.
The problem of the investor, however, is solved A feasible solution x* 僆 X is efficient if and only if
here as far as multiobjective optimization is con- there is a vector ⑀ * 僆 p of bounds such that x* is
cerned. In standard portfolio selection, return is an optimal solution to problem 3 for all l = 1, …, p.
measured as expected return of the portfolio, a lin- At first glance, the ⑀-constraint method appears
ear function, and risk is measured as variance, a to be superior to the weighted-sum method. This
concave function. All efficient portfolios are deter- is, however, not so. To prove that for efficient solu-
mined by the weighted-sum method and all irrele- tion x* there is some ⑀ such that x* is optimal to
vant alternatives (following the less is better problem 3 one only has to choose ⑀k = fk(x*). But,
assumption) are eliminated. He or she now needs of course, in practice that value is unknown

WINTER 2008 51
Articles

10 10
9 9
8 8
7 7
6 6
5 5
Y = f(X)
4 4
3 3
f1(x) 3.7
2 2
yI
1 1
0 0
0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10

Figure 4. The ⑀-Constraint Method (left) and the Compromise Programming Method (right).

because x* is unknown. Therefore, the method is the method behaves. Note that the weighted-sum
appropriate for checking whether or not some x* is method is a special case of problem 4: For q = 1,
efficient or not, but it cannot easily be used to problem 4 reduces to the minimization of the
derive an algorithm to solve nonconvex MOPs. weighted sum of the objective functions.
The third scalarization idea of minimizing the If x* is a unique optimal solution to problem 4 or if
distance to a reference point can be exemplified ␭ > 0 and x* is an optimal solution to problem 4
with the compromise programming method. The then x* is efficient.
principle is simple: Because there is no feasible If x* is an optimal solution to problem 5 and ␭ > 0
solution that minimizes all objectives simultane- then x* is weakly efficient.
ously (and thus achieves the ideal point), why not If x* is a unique optimal solution to problem 5 and
try and find a feasible solution x such that f(x) is as ␭ > 0 then x* is efficient.
close as possible to the ideal point yI? Closeness The use of yI makes it impossible to obtain com-
can be measured, for example, by a weighted q dis- pleteness results. However, the results about cor-
tance. A weighting vector ␭ ≥ 0 and an integer 1 ≤ rectness still hold if yI is replaced by a utopian
q < ∞ define the two compromise programming point yU. With that modification, Sawaragi,
problems Nakayama, and Tanino (1985) proved that optimal
⎧ ⎫⎪ solutions to problem 4 are properly efficient and
⎪⎪⎪⎛ p
1
q ⎞q ⎪
min ⎨⎜⎜∑ λ k ( f k ( x) − y k ) ⎟⎟ : x ∈ X ⎪⎬ and
I ⎟ (4) that YN is contained in the closure of the set of the

⎪⎪⎝ k=1 ⎟⎠ ⎪⎪ outcomes f(x) of all optimal solutions obtained
⎪⎩ ⎪⎭
(when varying parameters q and ␭ over all possible
values). Because this result is true without convex-
min { max λ ( f k ( x) − y kI ) : x ∈ X }.
k =1,, p k
(5) ity assumption, it is a generalization of the results
for the weighted-sum method to nonconvex prob-
These are illustrated in figure 4. Assuming that ␭ lems. Indeed, the value of q necessary to obtain an
= (1, 1) the feasible points of Y with minimal dis- efficient solution x* to the MOP as an optimal solu-
tance to yI lie on a circle (q = 2) or a square (q = ∞) tion to problem 4 can be interpreted as “degree of
around yI. The choice of q and ␭ determines how nonconvexity” of the problem.

52 AI MAGAZINE
Articles

So far, I have introduced the three fundamental sible solution x* 僆 X is an efficient solution to
ideas for scalarization: aggregation by weights, problem 6 if and only if there is a positive weight
conversion of objectives to constraints, and mini- vector ␭ > 0 such that ␭T Cx*  T Cx for all feasible
mizing distance to the ideal point. More generally, solutions x 僆 X.
one can write a scalarization method as min{s(f(x)) To turn this theoretical result into a practical
: r(f(x))  0, x 僆 X}, where s : p   and r : p  algorithm extensions of the simplex algorithm
t for some t. That is, s is a scalar valued function have been proposed, such as Steuer (1985). These
of the objectives to be minimized (such as a multiobjective simplex algorithms allow deter-
weighted sum, choosing one objective, the dis- mining all efficient solutions to problem 6 by
tance to the ideal point), subject to the original explicitly computing all efficient extreme points of
constraints of the problem (x 僆 X) and some addi- X (all other efficient solutions are convex combi-
tional constraints on the objective values. nations of those) and proceed in three phases.
Phase I: By solving a single-objective linear pro-
gram it can be determined whether the MOLP is
Multiobjective feasible. If so, a first feasible extreme point solution
Linear Programming is found, otherwise the algorithm ends.
Phase II: By solving another single-objective lin-
The more information is available about an MOP,
ear program it can be determined whether the
the better the chance of solving it, that is, finding
MOLP has any efficient solutions. If so, the LP pro-
its efficient solutions. As explained in the previous
vides a positive weight vector ␭ such that an opti-
section, (weakly, properly) efficient sets of convex
mal extreme point solution x* to the weighted sum
problems can be found with the weighted-sum
LP min{␭T Cx : Ax = b, x ≥ 0} is an efficient extreme
method. In this section, I will restrict this further
point solution to problem 6.
to linear problems and talk about algorithms to
Phase III: By modifying the rule for selecting
solve multiobjective linear programs.
entering variables in the simplex algorithm to
A bit more notation is needed first. In linear ver-
account for the multiobjective nature of the prob-
sions of problem 1 the objective function can be
lem, it is possible to explore all efficient extreme
written as f(x) = Cx with a p × n matrix C. Con-
point solutions to X starting from x*.
straints are (without loss of generality) equality
The number of efficient extreme point solutions
constraints Ax = b where A is a m × n matrix plus
to an MOLP grows rapidly with the number of
nonnegativity constraints x  0 on the variables.
objectives. Consequently, multiobjective simplex
Thus, a multiobjective linear program is
algorithms are rather inefficient. Moreover, many
min {Cx : Ax = b, x  0} . (6) points in X map to the same point in Y, so finding
The oncologist deals with a multiobjective lin- all efficient solutions implies finding all x with Cx
ear program. Designing a radiotherapy treatment = y for all nondominated points y. This is much
involves the selection of the intensity of radiation stronger requirement than in the single-objective
beams. Intensity can be modulated across a beam, case, where one is normally content with finding
modeled by decomposing a beam into hundreds of one (not all) optimal solutions. Thus it seems more
beam elements, each of which has its own variable appropriate to find the nondominated set and, for
intensity. The radiation dose deposited in the each nondominated point y, one efficient solution
patient body is modeled as a linear function of x with y = Cx.
intensity. The oncologist specifies a dose to be The radiotherapy treatment planning problem
delivered to the tumor and prescribes upper limits the oncologist tries to solve has thousands of vari-
of tolerable dose to healthy organs. Because these ables and possibly hundreds of thousands of con-
prescriptions are most often not simultaneously straints, but only three objectives. The oncologist
achievable, a multiobjective linear program can be wouldn’t make a decision based on the intensities
formulated to minimize the underdosing of the of beamlets, but rather on the dose deposited in
tumor and the overdosing of healthy organs (Shao the patient’s body. There seems to be benefit in
2008). All objectives and constraints are linear. solving MOLPs in objective rather than decision
A multiobjective linear program is of course con- space.
vex. Recall that weakly (properly) efficient solu- Harold Benson (1998) proposed an algorithm to
tions to convex MOPs are characterized by weight do just that. For ease of explanation I will assume
vectors ␭ ≥ 0 (respectively ␭ > 0), whereas efficient that Y is bounded and has dimension p. The algo-
solutions cannot be characterized. A fundamental rithm first constructs a polytope S0 defined by axes-
result of multiobjective linear programming states parallel hyperplanes and a supporting hyperplane
that all efficient solutions are properly efficient to Y such that Y is contained in S0. This polytope is
and can therefore be obtained by the weighted- updated iteratively until it is assured that the non-
sum method using positive weighting vectors. dominated points of Sk are identical to those of Y.
More precisely, Isermann (1974) proved that a fea- In iteration k, the algorithm first finds a vertex yk of

WINTER 2008 53
Articles

jective combinatorial optimization (MOCO) prob-


lems. Usually, the entries of A, b, C are integers.
Apart from the feasible sets X in decision space and
10 Y in objective space the convex hull of Y extended
by all points dominated by some y 僆 Y is of utmost
9 importance (see figure 5). Recall that for MOLPs
(for example, when relaxing the constraint of bina-
8
ry variables in problem 7) all efficient solutions are
7 supported. The binary variables in problem 7, how-
ever, destroy the convexity and nonsupported effi-
6
cient solutions exist. Furthermore, for some sup-
5 ported efficient solutions x*, Cx* is an extreme
point of the convex hull of Y. These are called
4 extreme supported efficient solutions. For all other
3 supported efficient solutions, Cx* is in the relative
interior of a face of the polyhedron. The nonsup-
2 ported efficient solutions are denoted XnE. For all
nonsupported efficient solutions x*, Cx* is in the
1
interior of the convex hull of Y.
0 Figure 5 shows the feasible set in objective space
0 1 2 3 4 5 6 7 8 9 10 for some MOCO problem as circles. The shaded
area is the convex hull of Y extended by all domi-
nated points. The problem has four supported
nondominated points, three of which are extreme
points. (5, 3) and (6, 2) are nonsupported non-
dominated points.
The convex hull of Y extended by all dominated
Figure 5. Nondominated Points for a MOCO Problem. points is a polyhedron, thus it has finitely many
facets and therefore finitely many weight vectors
(the normals to the facets) are sufficient to find all
supported nondominated points by solving
weighted-sum problems, that is, by solving finite-
ly many single-objective combinatorial optimiza-
Sk–1 that is not contained in Y. It then finds the tion problems. Hence finding supported efficient
boundary point sk on the line connecting yk with solutions is the same as solving several (possibly
an arbitrary but fixed point in the interior of Y. many) single-objective combinatorial optimiza-
Finally it finds a supporting hyperplane to Y con- tion problems.
taining the boundary point sk and adds this hyper- Note that because of the discrete nature of the
plane to the description of Sk-1 to obtain Sk. Thus, problem, all efficient solutions are properly effi-
in every iteration Sk becomes a better description of cient, thus this distinction, which is so important
Y. The oncologist has a set of outcomes correspon- in nonlinear multiobjective optimization, disap-
ding to efficient treatment plans at his or her dis- pears for MOCO problems.
posal. He or she needs to select one of those taking Efficient sets can also be classified according to
into account his or her clinical judgment of the the definition of (Hansen 1979). Efficient solutions
individual patient’s case. x1, x2 are equivalent if Cx1 = Cx2. A complete set of
efficient solutions is a subset X* of XE such that for
all y 僆 YN there is x 僆 X* with f(x) = y. A minimal
Multiobjective complete set contains no equivalent solutions,
Combinatorial Optimization whereas the maximal complete set XE contains all
equivalent solutions.
In this section I discuss MOPs that can be formu-
Multiobjective combinatorial optimization
lated with binary variables, linear constraints, and
problems are hard problems in terms of computa-
linear objectives
tional complexity, that is, they are usually NP-
min{f(x) = Cx : Ax = b, x 僆 {0, 1}n}. (7) hard, #P-hard, and there are problem instances
These problems are often used to model optimiza- with an exponential number of nondominated
tion over some combinatorial structures such as points and thus exponentially many efficient solu-
paths, trees, or tours in a graph; for example, the tions. This is the case even for problems for which
traveling salesman problem belongs to this prob- the single-objective version is easily solvable by
lem class. For this reason, they are called multiob- polynomial time algorithms such as shortest path,

54 AI MAGAZINE
Articles

assignment, spanning tree, and network flow prob- those that are complete involve solving hard
lems; see results and references in Ehrgott (2005b). scalarized problems (most prominently the ⑀-con-
Although the number of efficient solutions, and straint method). Ehrgott (2005a) proved that the
even the number of extreme supported nondomi- general scalarization is NP-hard. The difficulty
nated points, can be exponential in the size of the comes from the maximum term in problem 8 and
instance, numerical tests often show that the num- the constraints on objective values.
ber of nonsupported nondominated points grows The elastic constraint method of Ehrgott and
exponentially with instance size, whereas the Ryan (2003) allows a sort of compromise between
number of supported nondominated points grows completeness and computability. Like the ⑀-con-
polynomially with instance size. However, these straint method it uses constraints on p – 1 objec-
numbers do strongly depend on the numerical val- tives but makes these elastic by allowing them to
ues of C so that general statements cannot be be violated and penalizing the violation in the
made. In fact the influence of the values of the objective function.
entries in C on the difficulty of the problem is very Ehrgott and Ryan (2003) proved that the
poorly understood at this time. method is correct and complete. It comprises both
Next I will discuss some methods to solve the weighted-sum and the ⑀-constraint method as
MOCO problems. First, I will review aspects of special cases. Although the elastic constraint mod-
scalarization in the context of MOCO problems. el is NP-hard in general it is often solvable in prac-
Apart from correctness and completeness men- tice. Numerical tests in Ehrgott and Ryan (2003)
tioned earlier, two further properties of scalariza- indicate that its computational behavior is much
tion methods are relevant. Computability is con- better than that of the ⑀-constraint method
cerned with whether or not the scalarized problem because it “respects” the problem structure better
is harder than the single-objective version of the and “limits damage” done by adding constraints
MOCO problem at hand. This question can be on objective values.
addressed from a theoretical (for example, NP- In fact Ehrgott and Ryan (2003) studied the air-
completeness) and a practical perspective (that is, line crew scheduling problem. They used a set par-
the computational effort needed to solve the titioning formulation to model the constraints of
scalarized problem). Linearity refers to the objec- allocating one crew to each scheduled flight. In
tive and constraints of the scalarized problem. addition to a linear cost function they employed a
Because problem 7 has linear objectives and con- measure of robustness to calculate the potential
straints, it is generally desirable to preserve it to delays that can be caused during disruptions. The
avoid having to solve nonlinear integer programs. second objective is then to minimize the nonro-
The most common methods applied in MOCO bustness of the crew schedule. Whereas the ⑀-con-
are the weighted-sum method (2) (despite being straint scalarization could not be solved within rea-
unable to find any nonsupported efficient solu- sonable time, the elastic constraint scalarization
tions), the ⑀-constraint method (3) and the com- could be solved in the same time as the single-
promise programming method (5) with maximum objective problem. The airline manager can use
norm, also called the weighted Chebychev this method in two ways — to either compute sev-
method. Note that the compromise programming
eral schedules and evaluate their cost and per-
method with norms other than l1 and l∞ leads to
formance during disruptions, or to find a schedule
nonlinear objectives. It turns out that these as well
that maximizes robustness for a some approximate
as some other methods that retain linearity and do
allowable cost.
not introduce additional variables are special cases
There is no single best approach to solve MOCO
of a general formulation. Let s(f(x)) denote the
problems. Algorithms are generally problem spe-
function
cific and exploit the problem structure as much as
p
max ⎡ ν (c x − y R )⎤ + ∑ ⎡ λ (c x − y R )⎤. possible. Exact solution techniques can be broadly
k =1,, p ⎣⎢ k k k ⎦⎥ ⎣⎢ k k
k =1
k ⎦⎥ classified into three groups. First, the single-objec-
tive problem is polynomially solvable and a solu-
that is, the sum of the weighted maximum and tion algorithm can be extended to handle multiple
weighted average deviation from a reference point objectives. Second, the single-objective problem is
yR, where ck is the k-th row of C. This is used in polynomially solvable and efficient algorithm to
min{s(f(x)) : Cx ≤ ⑀, x 僆 X}. (8) generate feasible solutions to the single-objective
Although the previously mentioned scalariza- problem in order of increasing objective values
tions are correct and linear, there is a conflict exist. In this case the two phase method explained
between completeness and computability. Those below often works well. Third, if the single-objec-
scalarizations that lead to problems that are not tive problem is NP-hard more general integer pro-
harder than the single-objective MOCO problem gramming methods or heuristics are called for.
usually fail to find all efficient solutions (most Sometimes it is possible to adapt single-objective
prominently the weighted-sum method), whereas algorithms for the multiobjective case. The multi-

WINTER 2008 55
Articles

10 10
9 9
8 8
7 7
6 6
5 5
4 4
3 3
2 2
1 1
0 0
0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10

Figure 6. Phase 1 (left) and Phase 2 (right) of the Two Phase Method.

objective shortest path problem consists in finding until no new solutions are found. It finds at least
the efficient paths from node s to node t in a one efficient solution for each nondominated
directed graph with arc lengths being vectors. extreme point in objective space. In figure 6 the
Label correcting and label setting algorithms can lexicographically minimal points (1, 8) and (7, 1)
be applied to problems with any number of objec- define ␭ = (6, 7). The resulting weighted-sum prob-
tives noting that labels are vectors and that each lem identifies (3, 4) as nondominated point. Next,
node may have a set of labels that do not dominate points (1, 8) and (3, 4) define ␭ = (2, 1) and points
one another. Newly generated labels have to be (3, 4) and (7, 1) define ␭ = (3, 4). Neither of the two
compared with existing labels to eliminate any weighted-sum problems does yield any new solu-
dominated labels. The multiobjective spanning tions.
tree problem can be solved by generalizations of In phase two, nonsupported efficient solutions
Prim’s and Kruskal’s algorithms. are computed, in general by some enumerative
If a direct extension of a single-objective algo- methods. The set of nondominated extreme
rithm is not available one may think of applying points, which is known after phase 1, restricts the
the so-called two phase method first described by search area in objective space to triangles defined
(Ulungu and Teghem 1995), which is specific for by two consecutive nondominated extreme points
biobjective problems. It is based on the distinction as shown in figure 6. Nonsupported efficient solu-
between supported and nonsupported efficient tions are not optimal for any weighted-sum prob-
solutions and the fact that supported solutions are lem. Several ideas have been proposed to find
those that can be found by the weighted-sum them, but the best two phase algorithms available
method. Supported efficient solutions are comput- follow the idea of ranking solutions of single-
ed in phase one, often following a dichotomic objective (weighted-sum) problems. Nonsupported
scheme. First, two lexicographically optimal solu- efficient solutions can be expected to have good
tions are found ((1, 8) and (7, 1) in figure 6). Two weighted-sum objective values because they are
current solutions with consecutive points in objec- inside the triangles defined by optimal weighted-
tive space are used to find a new weighting vector sum solutions. Thus, they are second, third, or k
␭, finally the weighted-sum problem min{␭TCx : x best solutions to weighted-sum problems. For
僆 X} is solved. The procedure proceeds recursively many polynomially solvable combinatorial opti-

56 AI MAGAZINE
Articles

mization problems efficient ranking algorithms to Chankong, V., and Haimes, Y. 1983. Multiobjective Deci-
find k best solutions do exist. They can be used to sion Making Theory and Methodology. New York: Elsevier
enumerate solutions to the weighted-sum problem Science.
in order of their objective value until it is guaran- Ehrgott, M., and Ryan, D. 2003. Constructing Robust
teed that any further solution will be dominated. Crew Schedules with Bicriteria Optimization. Journal of
In figure 6, points (6, 2) and (5, 3) correspond to Multi-Criteria Decision Analysis 12(1): 139–150.
third and fourth-best solutions to the weighted- Ehrgott, M. 2005a. A Discussion of Scalarization Tech-
sum problem with ␭ = (3, 4) in the triangle given niques for Multiobjective Integer Programming. Annals of
Operations Research 147: 343–360.
by points (3, 4) and (7, 1), which both correspond
to optimal (first and second best solutions). The Ehrgott, M. 2005b. Multicriteria Optimization. Berlin:
Springer.
latter two points define ␭.
It is important to point out that the two phase Figueira, J.; Greco, S.; and Ehrgott, M., eds. 2005. Multiple
Criteria Decision Analysis: State of the Art Surveys. New
method always requires the solution of enumera-
York: Springer.
tion problems. Of course, in order to find a maxi-
Geoffrion, A. 1968. Proper Efficiency and the Theory of
mal complete set one must enumerate all solutions
Vector Maximization. Journal of Mathematical Analysis
x with Cx = y for all nondominated points y. But and Applications 22(3): 618–630.
even to find a minimal complete set enumeration
Hansen, P. 1979. Bicriterion Path Problems. In Multiple
is needed. For example, one can only guarantee to Criteria Decision Making Theory and Application, ed. G.
find an efficient solution corresponding to non- Fandel, and T. Gal, 109–127. Berlin: Springer Verlag.
dominated point (2, 6) in figure 6 by enumerating
Hartley, R. 1978. On Cone-Efficiency, Cone-Convexity
all optimal solutions to a weighted-sum problem and Cone-Compactness. SIAM Journal on Applied Mathe-
with weight vector (2, 1) normal to the line con- matics 34(2): 211–222.
necting (3, 4) and (1, 8). Isermann, H. 1974. Proper Efficiency and the Linear Vec-
To design a good two phase algorithm for a tor Maximum Problem. Operations Research 22(1): 189–
MOCO problem one needs efficient methods to 191.
find one and to enumerate all optimal solutions to Przybylski, A.; Gandibleux, X.; and Ehrgott, M. 2008.
as well as efficient algorithm to rank the feasible Two-Phase Algorithms for the Biobjective Assignment
solutions to the single-objective counterpart of the Problem. European Journal of Operational Research 185(2):
MOCO problem. Przybylski, Gandibleux, and 509–533.
Ehrgott (2008) give details of such an algorithm for Sawaragi, Y.; Nakayama, H.; and Tanino, T. 1985. Theory
the biobjective assignment problem. of Multiobjective Optimization. Orlando: Academic Press.
To solve MOCO problems for which the two- Shao, L. 2008. Multiobjective Linear Programming in
phase method is not applicable because of the lack Radiotherapy Treatment Planning. Ph.D. Dissertation,
of efficient single-objective optimization or rank- Department of Engineering Science, The University of
ing algorithms, one must resort to more general Auckland, Auckland, New Zealand.
integer programming techniques, such as branch Steuer, R. 1985. Multiple Criteria Optimization: Theory,
and bound, or heuristics. The branching part in Computation and Application. New York: John Wiley and
branch and bound algorithms can be done exactly Sons.
as in the single-objective case. The multiple objec- Ulungu, E., and Teghem, J. 1995. The Two-Phases
tives must be dealt with in the bounding step. Method: An Efficient Procedure to Solve Bi-Objective
Combinatorial Optimization Problems. Foundations of
Most methods use the ideal point of the problem
Computing and Decision Sciences 20(2): 149–165.
at a node of the branch and bound tree and elimi-
nate a node if that ideal point is dominated by a
feasible solution at another node. Such bounding Matthias Ehrgott is associate professor in operations
procedures can be very ineffective because the ide- research at the Department of Engineering Science, The
al point may be far away from the nondominated University of Auckland, New Zealand. He studied math-
points (see figure 5). A challenge here is the devel- ematics, economics, and computer science at the Uni-
opment of true multiobjective optimization versity of Kaiserslautern, Germany, obtaining his M.Sc.,
bounding schemes. Of course, today there are Ph.D., and Dr. habil. degrees in 1992, 1997, and 2001.
Ehrgott’s research interests are in integer programming
many MOCO problems that cannot be exactly
and multicriteria optimization and their application in
solved at all and heuristics need to be applied. That
real-world operations research problems. He has pub-
topic, however, is beyond the scope of this article. lished more than 50 refereed journal and proceedings
papers and written or edited several books and special
References issues of journals. Ehrgott is on the editorial board of
Benson, H. 1998. An Outer Approximation Algorithm for Management Science, OR Spectrum, Computers & OR, Asia
Generating All Efficient Extreme Points in the Outcome Pacific Journal of Operational Research, TOP, and INFOR.
Set of a Multiple Objective Linear Programming Problem. Currently he serves as vice president of the Operational
Journal of Global Optimization 13(1): 1–24. Research Society of New Zealand and as a member of the
Borwein, J. 1983. On the Existence of Pareto Efficient executive committee of the International Society on Mul-
Points. Mathematics of Operations Research 8(1): 64–73. tiple Criteria Decision Making.

WINTER 2008 57

You might also like