Professional Documents
Culture Documents
net/publication/220589717
How good are interior point methods? Klee-Minty. cubes tighten iteration-
complexity bounds
CITATIONS READS
27 557
3 authors:
Tamás Terlaky
Lehigh University
374 PUBLICATIONS 6,881 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Tamás Terlaky on 03 June 2014.
F U L L L E N G T H PA P E R
Received: 7 January 2005 / Revised: 16 August 2006 / Published online: 21 October 2006
© Springer-Verlag 2006
Abstract By refining a variant of the Klee–Minty example that forces the cen-
tral path to visit all the vertices of the Klee–Minty n-cube, we exhibit a nearly
worst-case example for path-following interior point methods. Namely, while
5
the theoretical iteration-complexity upper bound is O(2n n 2 ), we prove that
solving this n-dimensional linear optimization problem requires at least 2n − 1
iterations.
1 Introduction
While the simplex method, introduced by Dantzig [1], works very well in practice
for linear optimization problems, in 1972 Klee and Minty [6] gave an exam-
ple for which the simplex method takes an exponential number of iterations.
plexity lower bound. In this paper we show that a refined problem with the same
(2n ) iteration-complexity lower bound exhibits a nearly worst-case iteration-
5
complexity as the complexity upper bound is O(2n n 2 ). In other words, this new
N = O(2 n ), essentially √
example, with 2n 3 closes the iteration-complexity gap
with an ( ln3 N ) lower bound and an O( N ln N) upper bound.
N
min xn ,
subject to 0 ≤ x1 ≤ 1,
ε xk−1 ≤ xk ≤ 1 − ε xk−1 for k = 2, . . . , n.
How good are interior point methods? 3
The above minimization problem has 2n constraints, n variables and the feasible
region is an n-dimensional cube denoted by C. Some variants of the simplex
method take 2n − 1 iterations to solve this problem as they visit all the ver-
tices ordered by the decreasing value of the last coordinate xn starting from
v{n} = (0, . . . , 0, 1) till the optimal value x∗n = 0 is reached at the origin v∅ .
While adding a set h of redundant inequalities does not change the feasible
region, the analytic center χ h and the central path are affected by the addi-
tion of redundant constraints. We consider redundant inequalities induced by
hyperplanes parallel to the n facets of C containing the origin. The constraint
parallel to the facet H1 : x1 = 0 is added h1 times at a distance d1 and the
constraint parallel to the facet Hk : xk = εxk−1 is added hk times at a distance
dk for k = 2, . . . , n. The set h is denoted by the integer-vector h = (h1 , . . . , hn ),
d = (d1 , . . . , dn ), and the redundant linear optimization problem is defined by
min xn
subject to 0 ≤ x1 ≤ 1
ε xk−1 ≤ xk ≤ 1 − ε xk−1 for k = 2, . . . , n
0 ≤ d1 + x1 repeated h1 times
ε x1 ≤ d2 + x2 repeated h2 times
.. ..
. .
ε xn−1 ≤ dn + xn repeated hn times.
By analogy with the unit cube [0, 1]n , we denote the vertices of the Klee–Minty
cube C by using a subset S of {1, . . . , n}, see Fig. 1. For S ⊂ {1, . . . , n}, a vertex
vS of C is defined by
1, if 1 ∈ S
vS1 =
0, otherwise
1 − εvSk−1 , if k ∈ S
vSk = k = 2, . . . , n.
εvSk−1 , otherwise
ε= n
2(n+1) ,
d= n(2n+4 , . . . , 2n−k+5 , . . . , 25 ),
2n+8
h = 2 δn(n+1) 2n+7 (n+1)
, . . . , 2 δn(n+1)
n 2n+8 n+k−1
n−1 − δ n+k−2
4 A. Deza et al.
v {1,2}
v {1}
v∅
where 0 < δ ≤ 1
4(n+1) .
Note that we have: ε + δ < 12 ; that is, the δ-neighborhoods of the 2n vertices
are non-overlapping, and that h is, up to a floor operation, linearly dependent
on δ −1 . Proposition 2.1 states that, for Cnδ , the central path takes at least 2n − 2
turns before converging to the origin as it passes through the δ-neighborhood
of all the 2n vertices of the Klee-Minty n-cube; see Sect. 3.2 for the proof. Note
that the proof given in Sect. 3.2 yields a slightly stronger result than Propo-
sition 2.1: In addition to pass through the δ-neighborhood of all the vertices,
the central path is bent along the edge-path followed by the simplex method.
We set δ = 4(n+1)
1
in Propositions 2.3 and 2.4 in order to exhibit the sharpest
bounds. The corresponding linear optimization problem Cn1/4(n+1) depends only
on the dimension n.
Proposition 2.1 The central path P of Cnδ intersects the δ-neighborhood of each
vertex of the n-cube.
Since the number of iterations required by path-following interior point meth-
ods is at least the number of sharp turns, Proposition 2.1 yields a theoretical
lower bound for the iteration-complexity for solving this n-dimensional linear
optimization problem.
Corollary 2.2 For Cnδ , the iteration-complexity lower bound of path-following
interior point methods is (2n ).
Since the theoretical iteration-complexity
√ upper bound for path-following
interior point methods is O NL , where N and L respectively denote the
number of constraints and the bit-length of the input-data, we have:
Proposition 2.3 For Cn1/4(n+1) , the iteration-complexity upper bound of path-
3 11
following interior point methods is O(2n n 2 L); that is, O(23n n 2 ).
How good are interior point methods? 5
n+k
Proof We have N = 2n + nk=1 hk = 2n + nk=1 n2 22n+10 n+1 n − 2n+k+8
2k n+k
n+1
n and, since nk=1 n+1
n ≤ ne2 , we have N = O(22n n3 ) and L ≤
N ln d1 = O(22n n4 ).
Noticing that the only two vertices with last coordinates smaller than or equal
{1}
to εn−1 are v∅ and v{1} , with v∅n = 0 and vn = εn−1 , the stopping criterion can be
replaced by: stopping duality gap smaller than εn with the corresponding central
path parameter at the stopping point being µ∗ = εN . Additionally, one can check
n
Remark 2.5
(i) For Cn1/4(n+1) , by Corollary 2.2 and Proposition 2.4, the order of the iter-
ation-complexity of path-following interior
point methods is between 2n
5 √
and 2n n 2 or, equivalently, between lnN 3 N and N ln N.
(ii) The k-th coordinate of the vector d corresponds to the scalar d defined
in [2] for dimension n − k + 3.
(iii) Other settings for d and h ensuring that the central path visits all the
vertices of the Klee–Minty n-cube are possible. For example, d can be set
to (1.1, 22) in dimension 2.
(iv) Our results apply to path-following interior point methods but not to
other interior point methods such as Karmarkar’s original projective
algorithm [4].
Remark 2.6
(i) Megiddo and Schub [8] proved, for affine scaling trajectories, a result
with a similar flavor as our result for the central path, and noted that
their approach does not extend to projective scaling. They considered
the non-redundant Klee–Minty √ cube.
(ii) Todd and Ye [12] gave an ( 3 N) iteration-complexity lower bound
between two updates of the central path parameter µ.
(iii) Vavasis and Ye [13] provided an O(N 2 ) upper bound for the number of
approximately straight segments of the central path.
6 A. Deza et al.
(iv) A referee pointed out that a knapsack problem with proper objective
function yields an n-dimensional example with n + 1 constraints and n
sharp turns.
(v) Deza et al. [3] provided a non-redundant construction with N constraints
and N − 4 sharp turns.
we have Ah̃ ≥ 3b
2 .
Proof As ε = 2(n+1) n
and d = n(2n+4 , . . . , 2n−k+5 , . . . , 25 ), h̃ can be rewritten as
h̃ = 4δ d1 ( ε4n − 1ε ), . . . , εdk−1
k
( ε4n − ε1k ), . . . , εdn−1
n 3
ε n and Ah̃ ≥ 3b
2 can be rewritten
as
4 d1 4 1 4 4 1 6
− − − 2 ≥
δ d1 + 1 εn ε δ εn ε δ
4 4 1 4 2dk 4 1 4 4 1 6
− − + − − − k+1 ≥ for k = 2, . . . , n−1
δ εn ε δ dk + 1 εn εk δ εn ε δ
4 4 1 4 2dn 3 6
− − + ≥ ,
δ εn ε δ dn + 1 εn δ
which is equivalent to
1 1 3 4 1 3
− − d1 ≥ n − 2 + ,
ε2 ε 2 ε ε 2
1 2 1 3 8 1 1 3
− k+ − dk ≥ n − k+1 − + for k = 2, . . . , n − 1,
εk+1 ε ε 2 ε ε ε 2
How good are interior point methods? 7
2 1 3 4 1 3
+ − dn ≥ n − + .
ε n ε 2 ε ε 2
As ε12 − 1ε − 32 ≥ 12 , 1
ε − 3
2 ≥ 0, 1
ε2
− 3
2 ≥ 0 and 1
εk+1
+ 1
ε − 3
2 ≥ 0, the above
system is implied by
1 4
d1 ≥ n ,
2 ε
1 2 8
− k dk ≥ n for k = 2, . . . , n − 1,
εk+1 ε ε
2 4
dn ≥ n ,
ε n ε
n−k
1
as εk+1 − ε2k = 2
nεk
and 1
εn−k
= 2n−k 1 + n1 ≤ 2n−k+2 , the above system is
implied by
d1 ≥ 2n+5 ,
dk ≥ n2n−k+4 for k = 2, . . . , n − 1,
dn ≥ 2,
Corollary 3.2 With the same assumptions as in Lemma 3.1 and h = h̃, we have
Ah ≥ b.
h̃1 − h1 (h̃2 − h2 )ε 2
− ≤ ,
d1 + 1 d2 δ
h̃1 − h1 2(h̃k − hk )εk−1 (h̃k+1 − hk+1 )εk 2
− + − ≤ for k = 2, . . . , n − 1,
d1 dk + 1 dk+1 δ
h̃1 − h1 2(h̃n − hn )εn−1 2
− + ≤ ,
d1 dn + 1 δ
Corollary 3.3 With the same assumptions as in Lemma 3.1 and h = h̃, we have:
hk εk−1 hk+1 εk
dk +1 ≥ dk+1 + 4
δ for k = 1, . . . , n − 1.
Proof For k = 1, . . . , n − 1, one can easily check that the first k inequalities of
hk εk−1 hk+1 εk
Ah ≥ b imply dk +1 ≥ dk+1 + 4δ .
8 A. Deza et al.
The analytic center χ n = (ξ1n , . . . , ξnn ) of Cnδ is the unique solution to the
problem consisting of maximizing the product of the slack variables:
s1 = x1
sk = xk − εxk−1 for k = 2, . . . , n
s̄1 = 1 − x1
s̄k = 1 − εxk−1 − xk for k = 2, . . . , n
s̃1 = d1 + s1 repeated h1 times
.. ..
. .
s̃n = dn + sn repeated hn times.
n
max ln sk + ln s̄k + hk ln s̃k ,
x
k=1
n
max ln(xk − εxk−1 ) + ln(1 − εxk−1 − xk ) + hk ln(dk + xk − εxk−1 ) .
x
k=1
The optimality conditions (the gradient is equal to zero at optimality) for this
concave maximization problem give:
⎧ hk+1 ε
ε ε hk
⎪
⎨ σkn −
1
σk+1
n − 1
σ̄kn
− σ̄k+1
n + σ̃kn
−
σ̃k+1
n = 0 for k = 1, . . . , n − 1,
1
− σ̄ n + σ̃hnn
1
=0 (1)
⎪
⎩ σnn n n
σkn > 0, σ̄kn > 0, σ̃kn > 0 for k = 1, . . . , n,
where
σ1n = ξ1n
σkn = ξkn − εξk−1
n
for k = 2, . . . , n,
n n
σ̄1 = 1 − ξ1
σ̄kn = 1 − εξk−1
n
− ξkn for k = 2, . . . , n,
σ̃kn = dk + σkn for k = 1, . . . , n.
The following lemma states that, for Cnδ , the analytic center χ n belongs to the
neighborhood of the vertex v{n} = (0, . . . , 0, 1).
Proof Adding the nth equation of (1) multiplied by −εn−1 to the jth equation
of (1) multiplied by εj−1 for j = k, . . . , n − 1, we have, for k = 1, . . . , n − 1,
implying:
2hn εn−1 hk εk−1 εk−1 εk−1 2εn−1 εi n−2
εk−1
− = n − + +2 ≤ ,
σ̃nn σ̃k
n σk σ̄k
n σnn σ̄i+1
n σkn
i=k
h1 hk εk−1
which implies, since σ̃nn ≤ dn + 1, σ̃kn ≥ dk and d1 ≥ dk by Corollary 3.3,
The central path P of Cnδ can be defined as the set of analytic centers χ n (α) =
(xn1 , . . . , xnn−1 , α) of the intersection of the hyperplane Hα : xn = α with the
feasible region of Cnδ where 0 < α ≤ ξnn , see [9]. These intersections (α) are
called the level sets and χ n (α) is the solution of the following system:
ε
ε ε h
1
snk
− snk+1
− 1
s̄nk
− s̄nk+1
+ hs̃nk − s̃k+1
n =0 for k = 1, . . . , n − 1
k k+1 (2)
snk > 0, s̄nk > 0, s̃nk >0 for k = 1, . . . , n − 1,
where
sn1 = xn1
snk = xnk − εxnk−1 for k = 2, . . . , n − 1,
snn = α − εxn−1
s̄n1 = 1 − xn1
s̄nk = 1 − εxnk−1 − xnk for k = 2, . . . , n − 1,
s̄nn = 1 − α − εxnn−1
s̃nk = dk + snk for k = 1, . . . , n.
Lemma 3.5 For Cnδ , Cδk = {x ∈ C : s̄k ≥ εk−1 δ, sk ≥ εk−1 δ} and Ĉδk = {x ∈ C :
s̄k−1 ≤ εk−2 δ, sk−2 ≤ εk−3 δ, . . . , s1 ≤ δ}, we have: Cδk ∩ P ⊆ Ĉδk for k = 2, . . . , n.
10 A. Deza et al.
Proof Let x ∈ Cδk ∩ P. Adding the (k − 1)th equation of (2) multiplied by −εk−2
to the ith equation of (1) multiplied by εi−1 for i = j . . . , k − 2, we have, for
k = 2, . . . , n − 1,
which implies, since s̃nk−1 < dk−1 + 1, s̃nj > dj , s̃nk > dk and snk ≥ εk−1 δ and
s̄nk ≥ εk−1 δ as x ∈ Cδk ,
h1 hj εj−1
implying, since d1 ≥ dj by Corollary 3.3,
εk−2
that is, as 3δ ≤ − hd1 + dk−1 +1 − hkdε
2h k−1
by Corollary 3.2: snj ≤ εj−1 δ. Considering
1 k−1 k
the (k − 1)th equation of (2), we have
which implies, since s̃nk−1 < dk−1 + 1, s̃nk > dk and snk ≥ εk−1 δ and s̄nk ≥ εk−1 δ as
x ∈ Cδk ,
For k = 2, . . . , n, while Cδk , defined in Lemma 3.5, can be seen as the central part
of the cube C, the sets Tδk = {x ∈ C : s̄k ≤ εk−1 δ} and Bkδ = {x ∈ C : sk ≤ εk−1 δ},
How good are interior point methods? 11
T03
Ĉ03
B02 Ĉ02 T02
B03
A20 A30
Fig. 3 The sets A20 and A30 for the Klee–Minty 3-cube
can be seen, respectively, as the top and bottom part of C. Clearly, we have
C = Tδk ∪ Cδk ∪ Bkδ for each k = 2, . . . , n. Using the set Ĉδk defined in Lemma 3.5,
we consider the set Akδ = Tδk ∪ Ĉδk ∪ Bkδ for k = 2 . . . , n, and, for 0 < δ ≤ 4(n+1)
1
,
n
we show that the set Pδ = k=2 Aδ , see Fig. 2, contains the central path P. By
k
Lemma
n 3.4, the starting point χ n of P belongs to Nδ (v{n} ). Since P ⊂ C and
C = k=2 (Tδ ∪ Cδk ∪ Bkδ ), we have:
k
n
n
P=C∩P= Tδk ∪ Cδk ∪ Bkδ ∩ P = Tδk ∪ Cδk ∩ P ∪ Bkδ ∩ P,
k=2 k=2
n
n
P⊆ Tδk ∪ Ĉδk ∪ Bkδ = Akδ = Pδ
k=2 k=2
Remark that the sets Cδk , Ĉδk , Tδk , Bkδ and Akδ can be defined for δ = 0, see Fig. 3,
and that the corresponding set P0 = nk=2 Ak0 is precisely the path followed
by the simplex method on the original Klee-Minty problem as it pivots along
the edges of C. The set Pδ is a δ-sized (cross section) tube along the path P0 .
See Fig. 4 illustrating how P0 starts at v{n} , decreases with respect to the last
coordinate xn and ends at v∅ .
12 A. Deza et al.
v {1,3} v {2,3}
v {1,2,3}
v {2}
v {1,2}
∅
v v {1}
We consider the point x̄ of the central path which lies on the boundary of the
δ-neighborhood of the highest vertex v{n} . This point is defined by: s1 = δ, sk ≤
εk−1 δ for k = 2, . . . , n − 1 and s2n ≤ εn δ. Note that the notation sk for the cen-
tral path (perturbed complementarity) conditions, yk sk = µ for k = 1, . . . , pn ,
is consistent with the slacks introduced after Corollary 3.3 with sn+k = s̄k for
k = 1, . . . , n and spi +k = s̃k for k = 1, . . . , hi+1 Ê and i = 0, . . . , n − 1. Let µ̄
denote the central path parameter corresponding to x̄. In the following, we
prove that µ̄ ≤ εn−1 δ which implies that any point of the central path with
corresponding parameter µ ≥ µ̄ belong to the interior of the δ-neighborhood
of the highest vertex v{n} . In particular, it implies that by setting the central path
parameter to µ0 = 1, we obtain a starting point which belongs to the interior of
the δ-neighborhood of the vertex v{n} .
2n
n
pk
max z=− yk − dk yi
k=n+1 k=1 i=pk−1 +1
subject to yk − εyk+1 − yn+k − εyn+k+1
pk+1
pk
+ yi − ε yi = 0 for k = 1, . . . , n − 1
i=pk−1 +1 i=pk +1
pn
yn − y2n + yi = 1
i=pn−1 +1
yk ≥ 0 for k = 1, . . . , pn ,
2n+h
1
y1 − yn+1 − 2 εyn+2 + ε2 yn+3 + · · · + εn−1 y2n + yi = εn−1
i=2n+1
which implies
1
2n+h
n−1
2ε y2n ≤ y1 + yi
i=2n+1
µ̄
implying, since for i = 2n + 1, . . . , 2n + h1 , d1 ≤ si yields yi ≤ d1 , that
µ̄h1 µ̄ µ̄h1
2εn−1 y2n ≤ y1 + = + .
d1 δ d1
µ̄
Since for i = pn−1 + 1, . . . , pn , si = dn + x̄n − εx̄n−1 ≤ dn + 1 yields yi ≥ dn +1 ,
the last dual constraint implies
pn
µ̄hn
y2n ≥ yi − 1 ≥ −1
dn + 1
i=pn−1 +1
which, combined with the previously obtained inequality, gives µ̄ 2hdnnε+1 − hd1
n−1
1
1
µ̄ ≤ εn−1 δ.
Acknowledgments We would like to thank an associate editor and the referees for pointing out
the paper [8] and for helpful comments and corrections. Many thanks to Yinyu Ye for precious sug-
gestions and hints which triggered this work and for informing us about the papers [12, 13]. Research
supported by an NSERC Discovery grant, by a MITACS grant and by the Canada Research Chair
program.
References
1. Dantzig, G.B.: Maximization of a linear function of variables subject to linear inequalities. In:
Koopmans, T.C. (ed.) Activity Analysis of Production and Allocation, pp. 339–347. Wiley, New
York (1951)
2. Deza, A., Nematollahi, E., Peyghami, R., Terlaky, T.: The central path visits all the vertices of
the Klee–Minty cube. Optim. Methods Softw. 21–5, 851–865 (2006)
3. Deza, A., Terlaky, T., Zinchenko, Y.: Polytopes and arrangements: diameter and curvature.
AdvOL-Report 2006/09, McMaster University (2006)
4. Karmarkar, N.K.: A new polynomial-time algorithm for linear programming. Combinatorica
4, 373–395 (1984)
14 A. Deza et al.
5. Khachiyan, L.G.: A polynomial algorithm in linear programming. Soviet Math. Doklady 20,
191–194 (1979)
6. Klee, V., Minty, G.J.: How good is the simplex algorithm? In: Shisha, O. (ed.) Inequalities III,
pp. 159–175. Academic, New York (1972)
7. Megiddo, N.: Pathways to the optimal set in linear programming. In: Megiddo, N. (ed.) Progress
in Mathematical Programming: Interior-Point and Related Methods. Springer, Berlin Heidel-
berg New York pp. 131–158 (1988); also in: Proceedings of the 7th Mathematical Programming
Symposium of Japan, pp. 1–35. Nagoya, Japan (1986)
8. Megiddo, N., Shub, M.: Boundary behavior of interior point algorithms in linear programming.
Math. Oper. Res. 14–1, 97–146 (1989)
9. Roos, C., Terlaky, T., Vial, J.-Ph.: Theory and algorithms for linear optimization: an inte-
rior point approach. In: Wiley-Interscience Series in Discrete Mathematics and Optimization.
Wiley, New York (1997)
10. Sonnevend, G.: An “analytical centre” for polyhedrons and new classes of global algorithms
for linear (smooth, convex) programming. In: Prékopa, A., Szelezsán, J., Strazicky, B. (eds.)
System Modelling and Optimization: Proceedings of the 12th IFIP-Conference, Budapest 1985.
Lecture Notes in Control and Information Sciences, Springer, Berlin Heidelberg New York
Vol. 84, pp. 866–876 (1986)
11. Terlaky, T., Zhang, S.: Pivot rules for linear programming – a survey. Ann. Oper. Res. 46,
203–233 (1993)
12. Todd, M., Ye, Y.: A lower bound on the number of iterations of long-step and polynomial
interior-point linear programming algorithms. Ann. Oper. Res. 62, 233–252 (1996)
13. Vavasis, S., Ye, Y.: A primal-dual interior-point method whose running time depends only on
the constraint matrix. Math. Programm. 74, 79–120 (1996)
14. Wright, S.J.: Primal-Dual Interior-Point Methods. SIAM Publications, Philadelphia (1997)
15. Ye, Y.: Interior-Point Algorithms: Theory and Analysis. Wiley-Interscience Series in Discrete
Mathematics and Optimization. Wiley, New York (1997)