You are on page 1of 5

See discussions, stats, and author profiles for this publication at: https://www.researchgate.

net/publication/255648044

THE CONSTRAINED AFFINE PROJECTION ALGORITHM — DEVELOPMENT AND


CONVERGENCE ISSUES

Article · January 2000

CITATIONS READS

16 143

2 authors:

Marcello L. R. de Campos Jose Apolinario


Federal University of Rio de Janeiro Instituto Militar de Engenharia (IME)
135 PUBLICATIONS   1,313 CITATIONS    127 PUBLICATIONS   1,184 CITATIONS   

SEE PROFILE SEE PROFILE

Some of the authors of this publication are also working on these related projects:

Master od Science View project

D.Sc. Thesis: Algorithms for New Scenarios on Communications View project

All content following this page was uploaded by Jose Apolinario on 18 May 2017.

The user has requested enhancement of the downloaded file.


THE CONSTRAINED AFFINE PROJECTION ALGORITHM — DEVELOPMENT AND
CONVERGENCE ISSUES

Marcello L. R. de Campos1 and José Antonio Apolinário Jr.2

1 COPPE/Universidade Federal do Rio de Janeiro 2 Escuela Politécnica del Ejército


Programa de Engenharia Elétrica Facultad de Ingienerı́a Electrónica
P.O. Box 68504, Rio de Janeiro, RJ Valle de Los Chillos, Quito
21.945-970 Brazil Apartado 231-B Ecuador
campos@lps.ufrj.br apolin@espe.edu.ec

ABSTRACT 2. THE AFFINE-PROJECTION ALGORITHM


This paper introduces the constrained version of the Affine
Projection Algorithm. The new algorithm is suitable for The Affine Projection (AP) Algorithm updates its coeffi-
linearly-constrained minimum-variance applications, which cient vector such that the new solution belongs to the in-
include beamforming and multiuser detection for commu- tersection of L hyperplanes defined by the present and the
nications systems. The paper also discusses important as- L , 1 previous data pairs. The minimization problem used
pects of convergence and stability of constrained normal- to derive its updating formula is given by
ized adaptation algorithms in general. It is shown that nor-
malization may introduce bias in the final solution. Com- w(k + 1) = min k w , w(k ) k (1)
puter simulations are also included providing extra insight
w
to the algorithm performance. subjected to

1. INTRODUCTION d(k ) = XT (k )w (2)

Adaptation algorithms which satisfy linear constraints en- where


counter application in several areas of signal processing and
communications, such as beamforming, spectral estimation, d(k ) = [d(k ) d(k , 1)    d(k , L + 1)]T
multiuser detection for communication systems, etc. A ro-
bust algorithm which does not require reinitialization and X(k ) = [x(k ) x(k , 1)    x(k , L + 1)] (3)
incorporates the constraints into the solution was first intro-
duced by Frost [1]. More recently other constrained adap- where x(k ) = [x(k ) x(k , 1)    x(k , N )]T and w has
tation algorithms were introduced which are tailored to spe- order N or N + 1 elements.
cific applications or present advantageous performance re- The updating equations for the AP algorithm are ob-
garding convergence and robustness (see, e.g., [2][3]). tained from the solution of this minimization problem, and
The affine-projection (AP) algorithm is among the promi- they are presented in Table 1 [6][7]. A step-size () was
nent unconstrained adaptation algorithms that may have a used to control misadjustment and a small constant ( ) mul-
good compromise between fast convergence and low com- tiplied by the L  L identity matrix was used to improve
putational complexity. By adjusting the number of projec- robustness.
tions, or alternatively, the number of reuses, performance
can be controlled from that of the normalized least mean
squares (NLMS) algorithm to that of the sliding-window re- Table 1: The Affine Projection Algorithm
cursive least squares (RLS) algorithm [4][5]. AP Algorithm
In this article we develop and analyze the constrained for each k
version of the AP algorithm using the same framework al- f e(k) = d(k) , XT (k)w(k,)1
ready used for other normalized constrained algorithms, such t(k) = XT (k)X(k) + I e(k)
as the constrained NLMS and BNDR-LMS algorithms. We w(k + 1) = w(k) + X(k)t(k)
also show through analysis that normalization may intro-
g
duce bias.
3. THE CONSTRAINED AFFINE-PROJECTION and
ALGORITHM
e(k ) = d(k ) , XT (k )w(k ) (12)
In lineraly constrained adaptive filtering, the constraints are For L = 1 or L = 2, the above relations will result in
given by the following set of J equations the Constrained NLMS or Constrained BNDR-LMS algo-
rithms [8], respectively. For all constrained algorithms men-
CT w(k ) = f (4) tioned here, the simplification Pw(k ) + F = w(k ) should
where C is a (N +1)  J constraint matrix and f is a vector
be avoided in a finite precision environment, for accumula-
tion of round-off errors may cause the solution to drift away
containing the J constraint values.
from the constraint hyperplane [1].
Recalling the constrained LMS algorithm presented by
The equations of the Constrained Affine Projection Al-
gorithm are summarized in Table 2, where a step-size 0 <
Frost in [1] and realizing that it corresponds to the projec-
  1 and a small constant  were used.
tion of the unconstrained LMS solution onto the hyperplane
defined by (4), we may write

w(k + 1) = PwLMS (k + 1) + F Table 2: The Constrained Affine Projection Algorithm.


CAP Algorithm
= P[w(k) + e(k)x(k)] + F for each k
f e(k) = d(k) , XT (k)w(k),1
with
t(k) = XT (k) P X(k) + I e(k)
P = I , C(CT C),1 CT w(k + 1) = P [w(k) + X(k) t(k)] + F
(5)
g
the projection matrix (for a projection onto the homoge-
neous hyperplane defined by CT w(k ) = 0) and vector
4. ON THE CONVERGENCE OF THE CAP
F = C(CT C),1 f (6) ALGORITHM
used to move the projected solution back to the constraint For unconstrained adaptation algorithms, it is usually ex-
hyperplane. pected that convergence of the coefficients in the mean can
The following optimization approach was used to ob- be assured as the number of iterations goes to infinity. For
tain the constrained version of the Affine Projection (CAP) normalized algorithms, such as the NLMS, BNDR-LMS, or
Algorithm quasi-Newton [9] algorithms, convergence with probability
w(k + 1) = min k w , w(k ) k
one is usually more tractable and is sometimes preferred.
(7)
w As the CAP algorithm is a normalized algorithm, we will
favor the latter approach in the analysis to be presented in
subjected to this section.
( Let wo be the optimal solution to the constrained opti-
d(k ) , XT (k )w = 0
(8) mization problem, i.e.,
CT w = f , ,1 , T ,1 
wo = R,1 p , R,1 C CT R,1 C C R p ,f
We can now use Lagrange multipliers in the following (13)
and let d(k ) be modeled as
objective function to be minimized

 = [w , w(k)] [w , w(k)]
T
d(k ) = XT (k )wo (14)

+ T1 [d(k) , XT (k)w] If the coefficient-error vector is defined as


w(k) = w(k) , wo (15)
+ T2 [CT w , f ] (9)
then we can easily verify (after using the fact that the opti-
and the solution is the equation for the CAP algorithm: mal solution satisfies the constraints i.e., f , CT wo = 0),
that
w(k + 1) = P[w(k ) + X(k ) t(k )] + F (10)
w(k + 1) =
with
PfI , X(k )[XT (k )PX(k )],1 XT (k )gw(k )
t(k ) = [X (k ) P X(k )],1 e(k )
T
(11) (16)
−10
In order to guarantee convergence to zero everywhere
L=1 (NLMS)
[10] of the system described by the set of first-order ho- −20
L=2 (BNDR−LMS)
mogeneous equations above, we should ascertain that the L=3
−30 L=4
transition matrix is time-invariant with all its eigenvalues L=5
strictly inside the unit circle. This is clearly not satisfied. −40

Analysis of convergence in the mean requires independence −50

MSE (dB)
assumptions that are not fulfilled by (16). Furthermore, the
projection matrix P is not a Lyapunov transformation [11] −60

and cannot be used to define a stable equivalent system (in −70

the sense of Lyapunov) as was done in [12]. In fact, if we −80


rewrite (16) as
−90

w(k + 1) = PT(k)w(k) (17) −100


0 50 100 150 200 250 300 350 400 450 500
Number of iterations

we notice that the eigenvalues of T(k ) are not necessarily


inside the unit circle, although those of PT(k ) are always Figure 1: Learning curves for the CAP Algorithm.
either 1 or 0. Therefore, the constant projection matrix P
projects the drifting vector T(k )w(k ) onto the subspace 0.3

L=1
orthogonal to the subspace spanned by the constraint ma- L=2
trix C, such that the norm of w(k + 1) is never greater
0.25
L=3

than the norm of w(k ). This prevents divergence from the


constraint plane, but does not assure that kw(k )k ,! 0:
0.2

E[mean(W(k)−Wopt)]
Unbiasedness is, therefore, not guaranteed. 0.15

0.1
5. SIMULATIONS RESULTS
0.05

A first experiment was carried out in a system identifica-


tion problem where the filter coefficients were constrained 0

to preserve linear phase at every iteration. For this example


we made N = 10 and, in order to fulfill the linear phase −0.05
0 10 20 30 40 50 60 70 80 90 100
k
requirement, we made
2 3 Figure 2: First experiment: no bias in the coefficient vector.
IN=2
C=4 0T 5 (18)
,JN=2 A second experiment was done where the received sig-
nal consists of three sinusoids in white noise:
with J being a reversal matrix (an identity matrix with all
lines in reversed order), and x(k ) = sin(0:3k) + sin(0:325k)
f = [0    0]T (19) + sin(0:7k) + n(k) (20)

This didactic setup was employed to show the improve- where n(k) is white noise with power such that the SNR is
ment of the convergence speed when L is increased. Due 40dB . The filter is constrained to pass frequency compo-
to the symmetry of C and the fact that f is a null vector, nents of 0.1rad/s and 0.25rad/s undistorted which results in
more efficient structures could be used [13]. The input sig- the following constraint matrix and vector:
2 3
cos[(N , 1)0:2]
nal consists of zero mean unity variance colored noise with
eigenvalue spread around 2000 and the reference signal was 1 cos(0:2) :::

C T = 64
6 1 cos(0:5) ::: cos[(N , 1)0:5] 7
7
sin[(N , 1)0:2]
obtained after filtering the input by a linear-phase FIR fil-
1 sin(0:2) ::: 5
sin[(N , 1)0:5]
ter and adding an observation noise with variance equal to
1e , 10. Fig. 1 shows the learning curves for the CAP algo- 1 sin(0:5) :::
rithm with values of L varying from 1 to 5. It is also clear (21)
from this figure that the misadjustment increases with L.
Also for this first experiment, Fig. 2 shows that we have
no bias in the coefficient vector after convergence. F T = [1 1 0 0] (22)
The norm of the coefficient-error vector for values of L in the coefficient-error vector. The simulation results of two
from 1 to 3 is depicted in Fig. 3. From this figure we can experiments including both cases of biased and unbiased so-
realize that, although faster, the CAP algorithm presents an lutions supported the analysis claims and evaluated the per-
increasing misadjustment with L, specially when this num- formance of the proposed algorithm.
ber of projections is higher than 2 (this value corresponding
to the BNDR-LMS algorithm). 7. REFERENCES
2.5
[1] O. L. Frost III, “An algorithm for linearly constrained adap-
L=1 tive array processing,” Proceedings of the IEEE, vol. 60,
L=2
L=3 pp. 926–935, Aug. 1972.
2

[2] L. S. Resende, J. M. T. Romano, and M. G. Bellanger, “A


fast least-squares algorithm for linearly constrained adaptive
1.5 filtering,” IEEE Trans. Signal Processing, vol. 44, pp. 1168–
E[||W(k)−Wopt||]

1174, May 1996.

1
[3] M. L. R. de Campos, S. Werner, and J. A. Apolinário Jr.,
“Householder-transform constrained LMS algorithms with
reduced-rank updating,” in Proc. International Conference
0.5 of Acoustics, Speech, and Signal Processing, Phoenix, USA,
pp. 1857–1960, 1999.
[4] M. L.R. de Campos, P. S. R. Diniz, and J. A¿ Apolinário Jr.,
0
0 1000 2000 3000 4000 5000 6000
Number of iterations
7000 8000 9000 10000 “On normalized data-reusing and affine-projections algo-
rithms,” in Proc. IEEE International Conference on Elec-
Figure 3: Coefficient-vector deviation for the second exper- tronics, Circuits, and Systems, Pafos, Cyprus, 1999.
iment. [5] M. Montazeri and P. Duhamel, “A set of algorithms linking
NLMS and block RLS algorithms,” IEEE Trans. Signal Pro-
The reason for this behavior is found in Fig. 4 where are cessing, vol. 43, pp. 444–453, Feb. 1995.
plotted the curves corresponding to the mean value of the [6] K. Ozeki and T. Umeda, “An adaptive filtering algorithm
coefficient-error vector (averaged in 10 independent trials) using an orthogonal projection to an affine subspace and
for the three first values of L. its properties,” Electronics and Communications in Japan,
vol. 67-A, no. 5, pp. 19–27, 1984.
0.1

L=1 [7] S. L. Gay and S. Tavathia, “The fast affine projection al-
0.08
L=2 gorithm,” in Proc. International Conference of Acoustics,
L=3
0.06 Speech, and Signal Processing, Detroit, USA, pp. 3023–
0.04
3026, 1995.
[8] J. A. Apolinário Jr., S. Werner, and P. S. R. Diniz, “Con-
E[mean(Wopt−W(k))]

0.02
strained normalized adaptive filters for CDMA mobile com-
0
munications,” in Proc. European Signal Processing Confer-
−0.02 ence, vol. IV, Rhodes, Greece, pp. 2053–2056, 1998.
−0.04 [9] M. L. R. de Campos and A. Antoniou, “Analysis of a new
−0.06
quasi-Newton adaptive filtering algorithm,” in Proc. IEEE
International Conference on Electronics, Circuits, and Sys-
−0.08
tems, Rhodes, Greece, 1996.
−0.1
1 2 3 4 5 6 7 8 9 10 [10] A. Papoulis, Probability, Random Variables, and Stochastic
k 4
x 10
Processes. McGraw-Hill, 3rd ed., 1991.
[11] F. R. Gantmacher, The Theory of Matrices, vol. II. New York:
Figure 4: Second experiment: there is an increasing (with
Chelsea Publishing Company, 1959.
L) bias in the coefficient vector.
[12] M. L. R. de Campos and A. Antoniou, “A new quasi-Newton
adaptive filtering algorithm,” IEEE Trans. Circuits and Sys-
tems — Part II, vol. 44, pp. 924–934, Nov. 1997.
6. CONCLUSIONS [13] L. S. Resende, J. M. T. Romano, and M. G. Bellanger, “Sim-
plified FLS algorithm for linear phase adaptive filtering,”
In this paper, we have introduced the constrained version in Proc. European Signal Processing Conference, vol. III,
of the Affine Projection Algorithm. Through analysis, we Rodes, Greece, pp. 1237–1240, 1998.
have verified that this type of algorithm may introduce bias

View publication stats

You might also like