Professional Documents
Culture Documents
38ab PDF
38ab PDF
a
Universita di Roma \La Sapienza", Dipartimento di Informatica e Sistemistica,
Via Buonarroti 12, 00185 Roma, Italy. e-mail : soler@peano.dis.uniroma1.it
b
University of Hamburg, Institute of Applied Mathematics, Bundesstrasse 55,
D-20146 Hamburg, Germany. e-mail: kanzow@math.uni-hamburg.de
Abstract
A new algorithm for the solution of large-scale nonlinear complementarity prob-
lems is introduced. The algorithm is based on a nonsmooth equation reformulation
of the complementarity problem and on an inexact Levenberg-Marquardt-type algo-
rithm for its solution. Under mild assumptions, and requiring only the approximate
solution of a linear system at each iteration, the algorithm is shown to be both glob-
ally and superlinearly convergent, even on degenerate problems. Numerical results
for problems with up to 10000 variables are presented.
1 Introduction
Roma, Italy
p
'(a; b) := a + b ? a ? b:
2 2
(1)
@ (2)
'(xn ; Fn(x))
The function (1) has been introduced by Fischer [12] in 1992. Since then it
has interested several researchers in the elds of complementarity, constrained
optimization and variational inequality problems, see [13] for a survey of its
applications.
We propose an inexact Levenberg-Marquardt-type method for the solution of
the semismooth system (x) = 0. This method requires, at each iteration, the
approximate solution of a symmetric positive semidenite and solvable linear
system. We show that, under very mild assumptions, the algorithm is locally
superlinearly and even quadratically convergent to a solution of the nonlinear
complementarity problem. In order to globalize the local method we perform
an Armijo-type line search to minimize the natural merit function
(x) := (x)T (x);
1
2
2
which, in spite of the nonsmoothness of , turns out to be continuously dif-
ferentiable.
The new algorithm is based, on the one hand on the recently developed theory
for the solution of semismooth systems of equations by Newton-type methods,
see [30,29,22], and, on the other hand, on the properties of the functions
and that have also been the subject of recent research [10,5]. From the
theoretical point of view, the new method compares favourably with existing
methods. Numerical tests on problems with up to 10000 variables are reported
to show also its practical viability. As far as we are aware of, the algorithm
we propose here is the rst algorithm for nonlinear complementarity problems
that requires, at each iteration, only the inexact solution of a linear system
(which is an \easy" task) and can nevertheless guarantee fast convergence
even on degenerate problems.
The paper is organized as follows: In Section 2 we recall some recent results
on semismooth functions and prove some related new results that are needed
to analyze the behavior of the algorithm. Some important properties of the
operators and are summarized in Section 3. Section 4 contains a detailed
description of the new method along with the corresponding global and local
convergence theory. Some implementation issues of the algorithm are discussed
in Section 5, while the numerical tests are reported in Section 6.
A few words about notation. Given a function G : IRn ! IRn ; we denote by
Gi its ith component function. If G is continuously dierentiable, G0(x) is the
Jacobian matrix of G at a point x 2 IRn: If G is directionally dierentiable, we
denote the directional derivative of G at x 2 IRn in the direction d 2 IRn by
G0(x; d): The symbol k:k indicates the Euclidean vector norm or its associated
matrix norm. The index set f1; : : : ; ng is indicated by I . If M 2 IRnn is a
matrix with elements mij ; i; j 2 f1; : : : ; ng; and if K and J are subsets of
f1; : : : ; ng; then MKJ denotes the jK j jJ j submatrix of M consisting of the
elements mij ; i 2 K; j 2 J:
3
dierentiable almost everywhere. If we indicate by DG the set where G is
dierentiable, we can dene the B-subdierential of G at x [29] as
@B G(x) := klim G0(xk );
x 2DG
xk !x
4
Proposition 3 Suppose that x is a BD-regular solution of G(x) = 0, where
G : IRn ! IRn is semismooth. Then there exists a neighborhood
of x and a
constant c such that
kH ? k c;
1
8y 2
; 8H 2 @B G(y):
The following result plays a crucial role in the proof of our characterization
theorem for Q-quadratic convergence.
Proposition 4 Assume that G : IRn ! IRn is locally Lipschitzian and strongly
semismooth at x 2 IRn ; and directionally dierentiable in a neighborhood of x:
Then
lim sup kG(x + h) ? G(x) ? Hhk < 1: (5)
h!0;H 2@G(x+h) khk 2
5
where again H k 2 @B G(xk ) and dk := xk+1 ? xk :
!
kek k c kG(xk ) + H k dk k kdk k + kG(xk ) ? G(x ) ? H k ek k : (10)
+1 2
kek k 2
kdk k 2
kek k 2
kek k 2
by Lemma 8.2.3 in [6]. Using (11), condition (8) and Proposition 4, we imme-
diately obtain
k!1 2
6
We are now in the position to dene and analyze a Levenberg-Marquardt-type
method for the solution of a system of nonsmooth equations. Given a starting
vector x 2 IRn ; let
0
xk = xk + dk ;
+1
(12)
The assertion then follows from Theorem 5 (a). We rst note that by the
upper semicontinuity of the generalized Jacobian and by Proposition 3, there
exists a positive constant such that
k(H k )T k ; k(H k )?T k : (16)
7
By (14) we have
G(xk ) + H k dk = (H k )?T (rk ? k dk ): (17)
Hence, by (15), (17) and (16), to prove superlinear convergence we only need
to show that
k r k ? k dk k
lim kdk k = 0: (18)
k!1
lim kr k = 0:
k
(19)
k!1 kdk k
so that, by (16) and the assumption krk k = o(k(H k )T G(xk )k), we have kdk k =
O(k(H k)T G(xk )k), from which (19) follows.
(b) In order to prove the Q-quadratic convergence of the sequence fxk g to x ;
we only need to verify that
lim sup kG(x k)d+k kH d k < 1:
k k k
(20)
k!1 2
The assertion then follows from Theorem 5 (b). We can reason as in the
previous point to show that (20) is equivalent to
Always reasoning as on the previous point we also have that kdk k = O(k(H k)T G(xk )k).
So, taking into account that by assumption fk g = O(k(H k)T G(xk )k) and
krk k = O(k(H k)T G(xk )k ), (21) readily follows. 2
2
We stress that the condition krk k = o(k(H k )T G(xk )k) in Theorem 6 (a) can
be replaced by the condition krk k = o(kG(xk )k) without destroying the Q-
superlinear convergence of the sequence fxk g: This follows from the bounded-
ness of the sequence fxk g and the boundedness of the generalized Jacobian on
8
bounded sets. Similarly, the sequence fxk g remains Q-quadratically conver-
gent if we assume fk g = O(kG(xk )k) and krk k = O(kG(xk )k ) in Theorem
2
6 (b).
Remark 7 We note that it is easy to check, following the proofs of Theorems
5 and 6 that actually the following statement holds. If fxk g is any sequence
(no matter how generated) converging to x , then, under the assumptions of
Theorem 6 (a), if dk is generated according to (14), it holds that
lim k x k + dk ? x k
= 0:
k!1 kxk ? x k
tions.
In proving the superlinear convergence rate of our method we shall use the
following result, proved in [8].
Theorem 9 Let f : IRn ! IR be an SC function and suppose that we have a
1
Then, for any 2 (0; 1=2), there exists a k such that for every k k
f (xk + dk ) f (xk ) + rf (xk )T dk : (23)
3 Properties of and
In this section we recall a few properties of the two operators and which
are directly relevant to the convergence theory of the method to be described
in the next section.
9
In the rst theorem below we summarize the dierential properties of and
we shall need [10,20].
Theorem 10 Let F : IRn ! IRn be given.
(a) If F is a semismooth function, then is also semismooth.
(b) If F is twice continuously dierentiable, then is strongly semismooth.
(c) If F is continuously dierentiable, then is also continuously dieren-
tiable, and its gradient at a vector x 2 IRn is given by r (x) = H T (x);
where H can be an arbitrary element in @ (x):
(d) If all component functions Fi are SC 1 functions, then is also an SC 1
function.
We remark the fact that, although is nonsmooth, is continuously dier-
entiable. This fact will be crucial in the design of our algorithm.
The following matrix-classes will be used in the subsequent denitions and
results.
Denition 11 A matrix M 2 IRnn is called a
(a) P -matrix if det(MJJ ) > 0 for all subsets J I ;
(b) P0 -matrix if det(MJJ ) 0 for all subsets J I ;
(c) semimonotone matrix if, for all x 0; x 6= 0; there exists an index i 2 I
such that xi > 0 and xi[Mx]i 0:
Obviously, every P -matrix is a P -matrix. Furthermore, it can easily be ver-
0
ied that positive denite matrices are P -matrices and positive semidenite
matrices are P -matrices. Moreover it is known that the class of semimonotone
0
matrices includes the class of P -matrices, see [4].
0
10
Based on this denitionn, we can state the following result which is an imme-
diate consequence of [5, Theorem 5.4].
Theorem 13 Let x 2 IRn be a b-regular solution of NCP(F ) such that the
Schur-complement
F0 (x ) ? F
0 (x )F 0 (x )? F 0 (x )
1
Proof. Part (a) follows from Corollary 4.4 (a) in [5], whereas part (b) is a
consequence of Corollary 3.9.7 in [4] and Corollary 4.4 (c) in [5]. 2
As far as the authors are aware of, these conditions for stationary points being
solutions of NCP(F ) are the weakest conditions known for any equation-based
method or for any (constrained or unconstrained) minimization reformulation
of NCP(F ).
11
4 The Nonsmooth Inexact Levenberg-Marquardt Algorithm
Step 0: Set k = 0:
Step 1: (stopping criterion) If kr (xk )k "; stop.
Step 2: (search direction calculation) Select an element H k 2 @B (xk ):
Find an approximate solution dk 2 IRn of the system
((H k )T H k + k I )d = ?(H k )T (xk ); (24)
where k 0 is the Levenberg-Marquardt parameter. If the condition
r (xk )T dk ?kdk kp (25)
is not satised, set dk = ?r (xk ):
Step 3: (line search) Find the smallest ik 2 f0; 1; 2; : : :g such that
(xk + 2?ik dk ) (xk ) + 2?ik r (xk )T dk : (26)
Set xk = xk + 2?ik dk ; k
+1
k + 1 and go to Step 1.
Note that the choice k = 0 at each step is allowed by the above algorithm.
In this case, provided that H k is nonsingular, the equation (24) is equivalent
to the generalized Newton equation
H k d = ?(xk ):
The nonsmooth Newton method [29] for the solution of system (x) = 0
is therefore a particular case of our scheme. In Step 2 we say that we want
to solve the linear system (24) inexactly. By this we mean that the vector
12
dk 2 IRn satises
((H k )T H k + k I )dk = ?(H k )T (xk ) + rk = ?r (xk ) + rk (27)
(the second equality follows by Theorem 10 (c)) for some suitable residual
vector rk . Note however that, as usual in truncated schemes, the vector rk is
not xed beforehands. Instead an iterative solver is used to solve the linear
system (24), and this method is stopped when the norm of the residual rk is
smaller than a prexed accuracy (see the next two sections).
In what follows, as usual in analyzing the behaviour of algorithms, we shall
assume that " = 0 and that the algorithm produces an innite sequence of
points.
Theorem 15 Assume that the sequence fk g is bounded and that the sequence
of residual vectors frk g satises the condition krk k k kr (xk )k; where k
is a sequence of positive numbers converging to 0 and such that k < 1 for
every k: Then each accumulation point of the sequence fxk g generated by the
above procedure is a stationary point of :
Proof. First assume that there exists an innite set K such that dk = ?r (xk )
for all k 2 K: Then the assertion follows immediately from well-known results
(see, e.g., Proposition 1.16 in [1]). Hence we can assume without loss of gen-
erality that dk is always given by (27).
We now show that the sequence fdk g is uniformly gradient related to fxk g
according to the denition given in [1], i.e. we show that for every convergent
subsequence fxk gK for which
lim r (xk ) 6= 0
k!1;k2K
(28)
there holds
13
and therefore
kdk k k(kr (xk ) ? rk k :
H k )T H k + k I k (32)
Note that the denominator in (32) is nonzero since otherwise we would have
r (xk ) ? rk = 0 by (31) which, in turn, since krk k k kr (xk )k with
k < 1; would be possible only if kr (xk )k = 0; so that xk would be a
stationary point and the algorithm would have stopped.
Since fk g is bounded by assumption and the generalized Jacobian is upper
semicontinuous, there exists a constant > 0 such that
1
k(H k )T H k + k I k 1
(recall that the sequence fxk g is bounded since xk ! x ). Taking into account
krk k = o(kr (xk )k); we therefore get from (32)
kdk k (kr (xk ) ? rk k)= (kr (xk )k ? krk k)= kr (xk )k(33)
1 1 2
Relation (29) now readily follows from the fact that we are assuming that
the direction satises (25) with p > 2 while the gradient r (xk ) is bounded
on the convergent sequence fxk g. Consider (30). If it is not satised there
exists a subsequence fxk gK 0 of fxk gK for which limk!1;k2K 0 jr (xk )T dk j = 0.
This implies, by (25), that limk!1;k2K 0 kdk k = 0. In view of (33), this in turn
implies limk!1;k2K 0 kr (xk )k = 0, contradicting (28). The assertion of the
theorem now follows from [1, Proposition 1.8]. 2
Note that, under the additional assumptions of Theorem 14, we get from
Theorem 15 that any accumulation point of the sequence fxk g is a solution of
NCP(F ).
The next result gives a mild sucient condition for the whole sequence to
converge to a single point.
Theorem 16 Let the assumptions of Theorem 15 hold. Let fxk g be any se-
quence generated by the algorithm. If one of the accumulation points of fxk g;
say x ; is an isolated solution of NCP(F ), then the entire sequence fxk g con-
verges to x :
Proof. Let K be a subset of f1; 2; : : :g such that fxk gK ! x . Since fkr (xk )kgK !
14
0 we have, either because dk = ?r (xk ) or by (25), with p > 2, that
fdk gK ! 0. The thesis then follows by Theorem 2.3 in [9]. 2
r (xk )T dk ? kdk k
3
2
(34)
kdk k (dk )T (H k )T H k + k I dk kdk k :
1
2
2
2
(35)
We now obtain from (27), (33), (35), krk k k kr (xk )k and the Cauchy-
Schwarz inequality:
r (xk )T dk = ?(dk )T (H k )T H k + k I dk + (rk )T dk
? kdk k + krk k kdk k
1
2
15
? kdk k + k kr (xk )k kdk k
1
2
? kdk k + k kdk k = :
1
2 2
2
From this, inequality (34) follows immediately by taking into account that
fk g ! 0. Therefore, statement (b) follows from (34), fkdk kg ! 0 and the
fact that p > 2 in (25).
(c) By Theorem 10 (d), the merit function is an SC function. Statement
1
dk . The assertions then follow directly from Theorem 6, taking into account
Theorem 10 (a) and (b) and the fact that r (xk ) = (H k )T (xk ) by Theorem
10 (c). 2
5 Implementation Issues
16
a solution of the system in at most n iterations in exact arithmetic (see, e.g.,
[28] for the semidenite case). This means, in particular, that we can solve
system (24) as accurately as we want, i.e., that it is always possible to make
the residual vector rk as close to 0 as we like. Having chosen to solve system
(24) by the conjugate gradient method, we have to show how it is possible to
calculate the product of any vector v times the matrix (H k )T H k + k I . In turn,
it is clear that this is possible if we are able to calculate the product of any vec-
tor v by H k and by (H k )T . To this end we rst have to specify which element
H k 2 @B (xk ) we select at the k-th iteration. In [5] we showed that an element
of @B (x) can be obtained in the following way. Let := fi : xki = 0 = Fi(xk )g
be the set of \degenerate indices" and dene z 2 IRn to be a vector whose
components zi are 1 if i 2 and 0 otherwise. Then, the matrix H k dened by
H k = A(xk ) + B (xk )F 0(xk );
where A and B are n n diagonal matrices whose ith diagonal elements are
given, respectively, by
( xi ? 1 if i 62 ;
Aii(x) = k xi ;Fi zxi k ? 1 if i 2 ;
( ( ))
k zi ;rFi x T z k
( ( ) )
and by
8 Fi x
>
< ( )
? 1 if i 62 ;
Bii(x) = > k xir;FFii xx Tkz
( ( ))
: k zi ;rFi x T z k ? 1 if i 2 ;
(
( )
( ) )
belongs to @B (xk ). It is now easy to see that, storing in two vectors the di-
agonal elements of the matrices A and B and using the routines that calulate
the product of a vector v by the matrices F 0 and F 0T we can evaluate also
the product of v by H k and by (H k )T . Two situations can be distinguished. If
= ; (i.e. if is dierentiable at xk ), then the product Hv can be calculated
performing the product F 0(xk )v plus O(n) arithmetic operations. If, instead,
6= ; (i.e. if is not dierentiable at xk ), then we need an additional cal-
culation of the product F 0(xk )T z in order to evaluate the diagonal elements
of A(xk ) and B (xk ). As expected, if is not dierentiable at xk , then the
calculation of H k v is more costly than in the case in which is dierentiable
at x. Similar considerations can be made in the case of the product (H k )T v
and we omit them.
We nally discuss the linesearch used in the algorithm. We use a nonmonotone
linesearch, which can be viewed as an extension of (26). To motivate this
variant we rst recall that it has been often observed, in the eld of nonlinear
complementarity algorithms, that the linesearch test used to enforce global
17
convergence can lead to very small stepsizes; in turn this can bring to very slow
convergence and even to a numerical failure of the algorithm. To circumvent
this problem many heuristics have been used, see, e.g., [18,26]. We propose to
substitute the linesearch test (26) by the following nonmonotone linesearch:
(xk + 2?ik dk ) W + 2?ik r (xk )T dk ; (36)
and M k are nonnegative integers bounded above for any k (we suppose for
simplicity that k is large enough so that no negative superscripts can occur in
(37)). This kind of linesearch has been rst proposed in [15] and since then it
has proved very useful in the unconstrained minimization of smooth functions.
In particular it has already been used in the eld of nonlinear complementarity
problems in [5]. Adopting the same proof techniques used in [15], it is easy
to see that all the results described in the previous section still hold if we
substitute the linesearch (36) to (26) in the algorithm of Section 4. If at each
iteration we take W = (xk ); we obtain exactly the algorithm of Section 3,
however much more freedom is allowed by (36). In particular, according to
the most recent experiences in nonmonotone minimization algorithms [32], we
chose to keep W xed as long as the algorithm seems to make progress and to
vary it only if for a certain prexed number of consecutive steps the function
values (xk ) increase; the precise scheme is as follows.
- Set W = (x ) at the beginning of the algorithm;
0
6 Numerical Results
18
For all i = 1; : : : ; n set
(
? gi(x ) if i odd or i > n=2;
Fi(x) = ggi((xx)) ?
i gi(x ) + 1 otherwise
stopped the algorithm after 100 iterations if the former stopping criterion
was not met. We used a single set of parameters for all the runs of the prob-
lems, more precisely we set: = 10? , = 10? and p = 2:1. The conjugate
4 8
gradient algorithm was stopped when the norm of the residual is smaller than
(0:1=(k + 1))kr (xk )k, but we also set a limit of 200 iterations to the conju-
gate gradient phase. Finally we turn to the choice of k . We rst note that the
classical sophisticated choices indicated, e.g., in [6], are not suitable for large
scale problems. We used a very simple, and yet seemingly eective strategy. If
in the previous iteration the quotient kr (xk? )k=kdk? k is greaterpthan 250
1 1
and the norm of the residual k min(xk ; F (xk ))k is smaller than k(0:1 n) then
we set k = 1, otherwise we set k = 0. We note that the rst test is a rough
indicator that \something is going wrong" while the second test makes the
possibility of the perturbation (i.e. k > 0) more and more unlikely the more
the process progresses, so that the nal fast convergence rate is preserved.
We run the algorithm on the 17 test problems with dimensions n = 100,
n = 1000 and n = 10000. The results are summarized in Table 1. For each
test we report: the dimension of the problem (n), the number of iterations (it),
the total number of iterations of the conjugate gradient algorithm used to solve
the linear systems (minor it), the number of functions evaluations (F), the nal
value
pn, sooftothemake
merit function and the norm of the residual vector divided by
this stationarity measure independent of the dimension of the
problem. If convergence occured to a solution of NCP(F ) dierent from x ,
this is indicated by an after the iteration number. Note that this evenience is
possible because we do not have any guarantee that x is the only solution of
the problem generated. In any case we observed that, even when the solution
found diers from x, this is so only for very few components.
The overall results seem to indicate that, on this set of test problems, the
algorithm is fairly robust and capable of nding a solution to the problem with
a limited amount of work. We see that the algorithm is practically insensitive
to the dimension of the problem, which is a typical feature of Newton-type
19
Table 1
Results for degenerate problems. p
Prob. n it minor it. F Res/ n
100 9 139 10 .82d-9 .29d-5
1 1000 9 159 10 .31d-7 .56d-5
10000 10 166 11 .45d-8 .67d-6
100 4 6 8 .13d-9 .16d-5
2 1000 4 6 7 .20d-12 .20d-7
10000 4 6 7 .20d-11 .20d-7
100 F
3 1000 F
10000 F
100 10 82 11 .38d-9 .29d-5
4 1000 11 84 12 .31d-8 .27d-5
10000 11 81 12 .28d-9 .23d-6
100 31 1786 36 .45d-8 .93d-5
5 1000 19 1414 20 .45d-7 .90d-5
10000 21 1060 24 .46d-6 .92d-5
100 F
6 1000 F
10000 F
100 25 321 31 .22d-8 .67d-5
7 1000 26 330 29 .83d-8 .41d-7
10000 31 340 36 .41d-6 .91d-5
100 13 128 14 .83d-9 .23d-5
8 1000 15 144 16 .30d-8 .14d-5
10000 15 139 16 .46d-7 .30d-5
100 14 150 15 .41d-9 .25d-5
9 1000 14 101 15 .72d-8 .24d-5
10000 15 105 16 .91d-7 .24d-5
100 16 156 18 .47d-12 .97d-7
10 1000 10 24 11 .24d-10 .22d-6
10000 11 32 13 .26d-6 .72d-5
100 4 6 5 .26d-17 .23d-9
11 1000 4 6 5 .26d-16 .23d-8
10000 4 6 5 .26d-15 .23d-9
100 10 45 11 .16d-9 .18d-5
12 1000 10 42 11 .15d-8 .17d-5
10000 10 42 11 .15d-7 .17d-7
100 25 162 33 .70d-10 .63d-6
13 1000 22 133 30 .19d-7 .30d-5
10000 18 110 25 .83d-7 .64d-5
100 14 390 15 .16d-8 .56d-5
14 1000 14 407 15 .95d-8 .28d-5
10000 13 702 14 .11d-6 .38d-5
100 14 297 15 .60d-9 .25d-5
15 1000 14 417 15 .28d-7 .41d-5
10000 22 793 28 .54d-7 .19d-5
100 10 259 11 .81d-9 .40d-5
16 1000 100 18828 108 .32d-4 .24d-3
10000 100 18962 102 .75d-3 .12d-4
100 10 72 11 .15d-9 .17d-5
17 1000 10 70 11 .33d-8 .26d-5
10000 10 68 11 .40d-8 .90d-6
methods. Furthermore, in all cases in which convergence occurred, a fast nal
rate of convergence was observed, showing that the assumptions necessary to
establish the superlinear rate of convergence of the method are usually met
in practice. We now comment on the failures of the algorithm. The algorithm
fails on Problems 3 and 6 for n = 100, n = 1000 and n = 10000. The reason of
the failure is that in these cases convergence occurs towards stationary points
of the merit function that are not solutions of NCP(F ). There is not much
20
Table 2
Results for nondegenerate problems, n = 1000. p
Prob. it minor it. F Res/ n
1 6 65 7 .21d-7 .64d-5
2 4 6 7 .40d-12 .28d-7
3 F
4 14 70 15 .39d-8 .28d-5
5 19 648 21 .30d-7 .78d-5
6 F
7 25 329 27 .36d-11 .85d-7
8 14 86 15 .38d-9 .87d-6
9 17 112 18 .63d-9 .11d-5
10 9 20 10 .15d-7 .55d-5
11 3 4 4 .13d-7 .51d-5
12 6 18 7 .83d-12 .40d-7
13 14 67 20 .72d-11 .12d-6
14 100 3191 796 .79d-1 .12d-1
15 12 136 16 .11d-7 .47d-5
16 11 141 12 .43d-9 .92d-6
17 10 61 11 .39d-9 .89d-6
21
The only new observation we can make with respect to these results is that
they are overall very similar to the previous ones, showing that the algorithm
is seemingly insensitive to whether the solution towards which convergence
occurs is a dierentiable point of the system (x) = 0 or not.
Acknowledgement
We are grateful to M.A. Gomes-Ruggiero and J.M. Martnez who very kindly
provided us with the fortran codes of the test problems.
References
22
[10] F. Facchinei and J. Soares, A new merit function for nonlinear complementarity
problems and a related algorithm, DIS Technical Report 15.94, Universita di
Roma \La Sapienza" (Roma, Italy, 1994).
[11] M.C. Ferris and J.-S. Pang, Engineering and economic applications of
complementarity problems, Technical Report, Computer Sciences Department,
University of Wisconsin (Madison, WI, 1995).
[12] A. Fischer, A special Newton-type optimization method, Optimization 24 (1992)
269{284.
[13] A. Fischer, An NCP-function and its use for the solution of complementarity
problems, to appear in D.-Z. Du, L. Qi and R. Womersley (eds.),
Recent Advances in Nonsmooth Optimization (Kluwer Academic Publishers,
Dordrecht, 1995).
[14] S.A. Gabriel and J.-S. Pang, An inexact NE/SQP method for solving
the nonlinear complementarity problem, Computational Optimization and
Applications 1 (1992) 67{91.
[15] L. Grippo, F. Lampariello and S. Lucidi, A nonmonotone line search technique
for Newton's method, SIAM Journal on Numerical Analysis 23 (1986) 707{716.
[16] M.A. Gomes-Ruggiero, J.M. Martnez and S.A. Santos, Solving nonsmooth
equations by means of quasi-Newton methods with globalization. To appear
in D.-Z. Du, L. Qi and R. Womersley (eds.), Recent Advances in Nonsmooth
Optimization (Kluwer Academic Publishers, Dordrecht, 1995).
[17] P.T. Harker and J.-S. Pang, Finite-dimensional variational inequality and
nonlinear complementarity problems: a survey of theory, algorithms and
applications, Mathematical Programming 48 (1990) 161{220.
[18] P.T. Harker and B. Xiao, Newton's method for the nonlinear complementarity
problem: a B-dierentiable equation approach, Mathematical Programming 48
(1990) 339{357.
[19] G. Isac, Complementarity Problems , Lecture Notes in Mathematics 1528
(Springer-Verlag, Berlin, 1992).
[20] H. Jiang and L. Qi, A new nonsmooth equations approach to nonlinear
complementarities, Applied Mathematics Report 94/31, School of Mathematics,
University of New South Wales (Sydney, Australia, October 1994).
[21] L. Luksan, Inexact trust region method for large sparse systems of nonlinear
equations, Journal of Optimization Theory and Applications 81 (1994) 569{590.
[22] J.M. Martnez and L. Qi, Inexact Newton methods for solving nonsmooth
equations, Applied Mathematics Report 93/9, School of Mathematics,
University of New South Wales (Sydney, Australia, 1993, revised April 1994).
[23] R. Miin, Semismooth and semiconvex functions in constrained optimization,
SIAM Journal on Control and Optimization 15 (1977) 957{972.
23
[24] J.-S. Pang, Inexact Newton methods for the nonlinear complementarity
problem, Mathematical Programming 36 (1986) 54{71.
[25] J.-S. Pang, Complementarity problems, in: R. Horst and P.M. Pardalos (eds.),
Handbook of Global Optimization (Kluwer Academic Publishers, Dordrecht,
1994), 271{338.
[26] J.-S. Pang and S.A. Gabriel, NE/SQP: A robust algorithm for the nonlinear
complementarity problem, Mathematical Programming 60 (1993) 295{337.
[27] J.-S. Pang and L. Qi, Nonsmooth equations: motivation and algorithms, SIAM
Journal on Optimization 3 (1993) 443{465.
[28] B.N. Pshenichny and J.M Danilin, Numerical Methods in Extremal Problems
(Mir Publisher, Moscow, 1978).
[29] L. Qi, A convergence analysis of some algorithms for solving nonsmooth
equations, Mathematics of Operations Research 18 (1993) 227{244.
[30] L. Qi and J. Sun, A nonsmooth version of Newton's method, Mathematical
Programming 58 (1993) 353{368.
[31] S.M. Robinson, Generalized equations, in: A. Bachem, M. Groetschel and
B. Korte (eds.), Mathematical Programming: The State of the Art (Springer-
Verlag, Berlin, Germany, 1983), 346{367.
[32] Ph.L. Toint, A non-monotone trust-region algorithm for nonlinear optimization
subject to convex constraints, Report 92/94, Department of Mathematics,
FUNDP, Namur, Belgium, 1994.
24