Professional Documents
Culture Documents
Copyright c 2002
John Wiley & Sons Ltd, The Atrium, Southern Gate, Chichester, West Sussex PO19 8SQ, England Telephone (C44) 1243 779777
Email (for orders and customer service enquiries): cs-books@wiley.co.uk Visit our Home Page on www.wileyeurope.com or www.wiley.com Reprinted with corrections March 2003 All Rights Reserved. No part of this publication may be reproduced, stored in a retrieval system or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, scanning or otherwise, except under the terms of the Copyright, Designs and Patents Act 1988 or under the terms of a licence issued by the Copyright Licensing Agency Ltd, 90 Tottenham Court Road, London W1T 4LP, UK, without the permission in writing of the Publisher. Requests to the Publisher should be addressed to the Permissions Department, John Wiley & Sons Ltd, The Atrium, Southern Gate, Chichester, West Sussex PO19 8SQ, England, or emailed to permreq@wiley.co.uk, or faxed to (C44) 1243 770571. Neither the author(s) nor John Wiley & Sons, Ltd accept any responsibility or liability for loss or damage occasioned to any person or property through using the material, instructions methods or ideas contained herein, or acting or refraining from acting as a result of such use. The author(s) and Publisher expressly disclaim all implied warranties, including merchantability of tness for any particular purpose. Designations used by companies to distinguish their products are often claimed as trademarks. In all instances where John Wiley & Sons is aware of a claim, the product names appear in initial capital or capital letters. Readers, however, should contact the appropriate companies for more complete information regarding trademarks and registration.
Other Wiley Editorial Ofces John Wiley & Sons Inc., 111 River Street, Hoboken, NJ 07030, USA Jossey-Bass, 989 Market Street, San Francisco, CA 94103-1741, USA Wiley-VCH Verlag GmbH, Boschstr. 12, D-69469 Weinheim, Germany John Wiley & Sons Australia Ltd, 33 Park Road, Milton, Queensland 4064, Australia John Wiley & Sons (Asia) Pte Ltd, 2 Clementi Loop #02-01, Jin Xing Distripark, Singapore 129809 John Wiley & Sons Canada Ltd, 22 Worcester Road, Etobicoke, Ontario, Canada M9W 1L1 Wiley also publishes its books in a variety of electronic formats. Some content that appears in print may not be available in electronic books.
British Library Cataloguing in Publication Data A catalogue record for this book is available from the British Library ISBN 0-470-84389-6
A Produced from L TEX les supplied by the authors, processed by Laserwords Private Limited, Chennai, India Printed and bound in Great Britain by Biddles Ltd, Guildford and Kings Lynn This book is printed on acid-free paper responsibly manufactured from sustainable forestry in which at least two trees are planted for each one used for paper production.
Contents
Preface Acknowledgements 1 Elements of signal theory 1.1 Signal space : : : : : : : : : : : : : : : : : : : : : : : : : : : : : Properties of a linear space : : : : : : : : : : : : : : : : Inner product : : : : : : : : : : : : : : : : : : : : : : : 1.2 Discrete signal representation : : : : : : : : : : : : : : : : : : : : The principle of orthogonality : : : : : : : : : : : : : : Signal representation : : : : : : : : : : : : : : : : : : : GramSchmidt orthonormalization procedure : : : : : : 1.3 Continuous-time linear systems : : : : : : : : : : : : : : : : : : : 1.4 Discrete-time linear systems : : : : : : : : : : : : : : : : : : : : : Discrete Fourier transform (DFT) : : : : : : : : : : : : The DFT operator : : : : : : : : : : : : : : : : : : : : : Circular and linear convolution via DFT : : : : : : : : : Convolution by the overlap-save method : : : : : : : : : IIR and FIR lters : : : : : : : : : : : : : : : : : : : : 1.5 Signal bandwidth : : : : : : : : : : : : : : : : : : : : : : : : : : The sampling theorem : : : : : : : : : : : : : : : : : : Heaviside conditions for the absence of signal distortion 1.6 Passband signals : : : : : : : : : : : : : : : : : : : : : : : : : : : Complex representation : : : : : : : : : : : : : : : : : : Relation between x and x .bb/ : : : : : : : : : : : : : : : Baseband equivalent of a transformation : : : : : : : : : Envelope and instantaneous phase and frequency : : : : 1.7 Second-order analysis of random processes : : : : : : : : : : : : : 1.7.1 Correlation : : : : : : : : : : : : : : : : : : : : : : : : : : Properties of the autocorrelation function : : : : : : : : 1.7.2 Power spectral density : : : : : : : : : : : : : : : : : : : : Spectral lines in the PSD : : : : : : : : : : : : : : : : : Cross-power spectral density : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
xxix xxxi 1 1 1 3 4 6 6 8 13 17 19 20 21 23 25 28 30 32 33 33 34 42 43 44 45 46 46 47 48
viii
Contents
1.8
1.12
1.13
Properties of the PSD : : : : : : : : : : : : : : : PSD of processes through linear transformations : PSD of processes through ltering : : : : : : : : 1.7.3 PSD of discrete-time random processes : : : : : : : Spectral lines in the PSD : : : : : : : : : : : : : PSD of processes through ltering : : : : : : : : Minimum-phase spectral factorization : : : : : : 1.7.4 PSD of passband processes : : : : : : : : : : : : : PSD of the quadrature components of a random process : : : : : : : : : Cyclostationary processes : : : : : : : : : : : : : The autocorrelation matrix : : : : : : : : : : : : : : : : : : Denition : : : : : : : : : : : : : : : : : : : : : Properties : : : : : : : : : : : : : : : : : : : : : Eigenvalues : : : : : : : : : : : : : : : : : : : : Other properties : : : : : : : : : : : : : : : : : : Eigenvalue analysis for Hermitian matrices : : : Examples of random processes : : : : : : : : : : : : : : : Matched lter : : : : : : : : : : : : : : : : : : : : : : : : Matched lter in the presence of white noise : : Ergodic random processes : : : : : : : : : : : : : : : : : : 1.11.1 Mean value estimators : : : : : : : : : : : : : : : : Rectangular window : : : : : : : : : : : : : : : Exponential lter : : : : : : : : : : : : : : : : : General window : : : : : : : : : : : : : : : : : : 1.11.2 Correlation estimators : : : : : : : : : : : : : : : : Unbiased estimate : : : : : : : : : : : : : : : : : Biased estimate : : : : : : : : : : : : : : : : : : 1.11.3 Power spectral density estimators : : : : : : : : : : Periodogram or instantaneous spectrum : : : : : Welch periodogram : : : : : : : : : : : : : : : : Blackman and Tukey correlogram : : : : : : : : Windowing and window closing : : : : : : : : : Parametric models of random processes : : : : : : : : : : : ARMA. p; q/ model : : : : : : : : : : : : : : : MA(q) model : : : : : : : : : : : : : : : : : : : AR(N ) model : : : : : : : : : : : : : : : : : : : Spectral factorization of an AR(N ) model : : : : Whitening lter : : : : : : : : : : : : : : : : : : Relation between ARMA, MA and AR models : 1.12.1 Autocorrelation of AR processes : : : : : : : : : : 1.12.2 Spectral estimation of an AR.N / process : : : : : : Some useful relations : : : : : : : : : : : : : : : AR model of sinusoidal processes : : : : : : : : Guide to the bibliography : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
48 49 50 50 51 52 53 54 54 56 63 63 63 63 64 65 67 73 74 76 78 80 81 82 82 82 83 84 84 85 86 86 90 90 91 91 94 94 94 96 98 99 101 102
Contents
ix
Bibliography : : : : : : : : : : : : : : : : Appendices : : : : : : : : : : : : : : : : 1.A Multirate systems : : : : : : : : : : 1.A.1 Fundamentals : : : : : : : : 1.A.2 Decimation : : : : : : : : : 1.A.3 Interpolation : : : : : : : : : 1.A.4 Decimator lter : : : : : : : 1.A.5 Interpolator lter : : : : : : : 1.A.6 Rate conversion : : : : : : : 1.A.7 Time interpolation : : : : : : Linear interpolation : : : : Quadratic interpolation : : 1.A.8 The noble identities : : : : : 1.A.9 The polyphase representation Efcient implementations : 1.B Generation of Gaussian noise : : : :
: : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
103 104 104 104 106 109 110 112 113 116 116 118 118 119 120 127 129 129 130 132 134 135 135 136 140 140 141 142 143 144 145 146 147 148 149 149 150 150 151 152 152 153 154 155 157
2 The Wiener lter and linear prediction 2.1 The Wiener lter : : : : : : : : : : : : : : : : : : : : : : : : Matrix formulation : : : : : : : : : : : : : : : : : Determination of the optimum lter coefcients : : The principle of orthogonality : : : : : : : : : : : Expression of the minimum mean-square error : : : Characterization of the cost function surface : : : : The Wiener lter in the z-domain : : : : : : : : : 2.2 Linear prediction : : : : : : : : : : : : : : : : : : : : : : : : Forward linear predictor : : : : : : : : : : : : : : Optimum predictor coefcients : : : : : : : : : : : Forward prediction error lter : : : : : : : : : : Relation between linear prediction and AR models First and second order solutions : : : : : : : : : : 2.2.1 The LevinsonDurbin algorithm : : : : : : : : : : : Lattice lters : : : : : : : : : : : : : : : : : : : : 2.2.2 The DelsarteGenin algorithm : : : : : : : : : : : : 2.3 The least squares (LS) method : : : : : : : : : : : : : : : : Data windowing : : : : : : : : : : : : : : : : : : : Matrix formulation : : : : : : : : : : : : : : : : : Correlation matrix : : : : : : : : : : : : : : : : Determination of the optimum lter coefcients : : 2.3.1 The principle of orthogonality : : : : : : : : : : : : : Expressions of the minimum cost function : : : : : The normal equation using the T matrix : : : : : : Geometric interpretation: the projection operator : : 2.3.2 Solutions to the LS problem : : : : : : : : : : : : : Singular value decomposition of T : : : : : : : : : Minimum norm solution : : : : : : : : : : : : : :
Contents
Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : Appendices : : : : : : : : : : : : : : : : : : : : : : : : : : 2.A The estimation problem : : : : : : : : : : : : : : : : : The estimation problem for random variables MMSE estimation : : : : : : : : : : : : : : : Extension to multiple observations : : : : : : MMSE linear estimation : : : : : : : : : : : MMSE linear estimation for random vectors : 3
: : : : : : : :
: : : : : : : :
: : : : : : : :
: : : : : : : :
: : : : : : : :
: : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
158 159 159 159 159 160 161 162 165 166 166 168 169 170 171 173 173 175 175 175 177 178 179 180 183 184 186 186 187 187 188 189 189 191 191 197 198 199 201 202 203 203 203 204 205
Adaptive transversal lters 3.1 Adaptive transversal lter: MSE criterion : : : : : : : : : : : : : : 3.1.1 Steepest descent or gradient algorithm : : : : : : : : : : : Stability of the steepest descent algorithm : : : : : : : : Conditions for convergence : : : : : : : : : : : : : : : : Choice of the adaptation gain for fastest convergence : : Transient behavior of the MSE : : : : : : : : : : : : : : 3.1.2 The least mean-square (LMS) algorithm : : : : : : : : : : Implementation : : : : : : : : : : : : : : : : : : : : : : Computational complexity : : : : : : : : : : : : : : : : Canonical model : : : : : : : : : : : : : : : : : : : : : Conditions for convergence : : : : : : : : : : : : : : : : 3.1.3 Convergence analysis of the LMS algorithm : : : : : : : : Convergence of the mean : : : : : : : : : : : : : : : : : Convergence in the mean-square sense (real scalar case) Convergence in the mean-square sense (general case) : : Basic results : : : : : : : : : : : : : : : : : : : : : : : : Observations : : : : : : : : : : : : : : : : : : : : : : : Final remarks : : : : : : : : : : : : : : : : : : : : : : : 3.1.4 Other versions of the LMS algorithm : : : : : : : : : : : : Leaky LMS : : : : : : : : : : : : : : : : : : : : : : : : Sign algorithm : : : : : : : : : : : : : : : : : : : : : : Sigmoidal algorithm : : : : : : : : : : : : : : : : : : : Normalized LMS : : : : : : : : : : : : : : : : : : : : : Variable adaptation gain : : : : : : : : : : : : : : : : : LMS for lattice lters : : : : : : : : : : : : : : : : : : : 3.1.5 Example of application: the predictor : : : : : : : : : : : : 3.2 The recursive least squares (RLS) algorithm : : : : : : : : : : : : Normal equation : : : : : : : : : : : : : : : : : : : : : Derivation of the RLS algorithm : : : : : : : : : : : : : Initialization of the RLS algorithm : : : : : : : : : : : : Recursive form of E min : : : : : : : : : : : : : : : : : : Convergence of the RLS algorithm : : : : : : : : : : : : Computational complexity of the RLS algorithm : : : : : Example of application: the predictor : : : : : : : : : : 3.3 Fast recursive algorithms : : : : : : : : : : : : : : : : : : : : : : 3.3.1 Comparison of the various algorithms : : : : : : : : : : :
Contents
xi
Block adaptive algorithms in the frequency domain : : : : : : : : : 3.4.1 Block LMS algorithm in the frequency domain: the basic scheme : : : : : : : : : : : : : : : : : : : : : : : : Computational complexity of the block LMS algorithm via FFT : : : : : : : : : : : : : : : 3.4.2 Block LMS algorithm in the frequency domain: the FLMS algorithm : : : : : : : : : : : : : : : : : : : : : : Computational complexity of the FLMS algorithm : : : : : Convergence in the mean of the coefcients for the FLMS algorithm : : : : : : : : : : : : 3.5 LMS algorithm in a transformed domain : : : : : : : : : : : : : : : 3.5.1 Basic scheme : : : : : : : : : : : : : : : : : : : : : : : : : On the speed of convergence : : : : : : : : : : : : : : : : 3.5.2 Normalized FLMS algorithm : : : : : : : : : : : : : : : : : 3.5.3 LMS algorithm in the frequency domain : : : : : : : : : : : 3.5.4 LMS algorithm in the DCT domain : : : : : : : : : : : : : : 3.5.5 General observations : : : : : : : : : : : : : : : : : : : : : 3.6 Examples of application : : : : : : : : : : : : : : : : : : : : : : : 3.6.1 System identication : : : : : : : : : : : : : : : : : : : : : Linear case : : : : : : : : : : : : : : : : : : : : : : : : : Finite alphabet case : : : : : : : : : : : : : : : : : : : : : 3.6.2 Adaptive cancellation of interfering signals : : : : : : : : : : General solution : : : : : : : : : : : : : : : : : : : : : : : 3.6.3 Cancellation of a sinusoidal interferer with known frequency 3.6.4 Disturbance cancellation for speech signals : : : : : : : : : : 3.6.5 Echo cancellation in subscriber loops : : : : : : : : : : : : : 3.6.6 Adaptive antenna arrays : : : : : : : : : : : : : : : : : : : : 3.6.7 Cancellation of a periodic interfering signal : : : : : : : : : Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : Appendices : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 3.A PN sequences : : : : : : : : : : : : : : : : : : : : : : : : : : : : : Maximal-length sequences : : : : : : : : : : : : : : : : : CAZAC sequences : : : : : : : : : : : : : : : : : : : : : Gold sequences : : : : : : : : : : : : : : : : : : : : : : : 3.B Identication of a FIR system by PN sequences : : : : : : : : : : : 3.B.1 Correlation method : : : : : : : : : : : : : : : : : : : : : : Signal-to-estimation error ratio : : : : : : : : : : : : : : : 3.B.2 Methods in the frequency domain : : : : : : : : : : : : : : : System identication in the absence of noise : : : : : : : : System identication in the presence of noise : : : : : : : 3.B.3 The LS method : : : : : : : : : : : : : : : : : : : : : : : : Formulation using the data matrix : : : : : : : : : : : : : Computation of the signal-to-estimation error ratio : : : : 3.B.4 The LMMSE method : : : : : : : : : : : : : : : : : : : : : 3.B.5 Identication of a continuous-time system : : : : : : : : : : 3.4
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
205 206 206 207 209 211 211 212 214 214 214 215 216 216 216 217 220 221 222 224 224 225 226 227 229 233 233 233 235 236 239 239 241 242 242 243 244 246 246 249 251
xii
Contents
Transmission media 4.1 Electrical characterization of a transmission system : : : : : Simplied scheme of a transmission system : : : Characterization of an active device : : : : : : : Conditions for the absence of signal distortion : : Characterization of a 2-port network : : : : : : : Measurement of signal power : : : : : : : : : : 4.2 Noise generated by electrical devices and networks : : : : : Thermal noise : : : : : : : : : : : : : : : : : : : Shot noise : : : : : : : : : : : : : : : : : : : : : Noise in diodes and transistors : : : : : : : : : : Noise temperature of a two-terminal device : : : Noise temperature of a 2-port network : : : : : : Equivalent-noise models : : : : : : : : : : : : : Noise gure of a 2-port network : : : : : : : : : Cascade of 2-port networks : : : : : : : : : : : : 4.3 Signal-to-noise ratio (SNR) : : : : : : : : : : : : : : : : : SNR for a two-terminal device : : : : : : : : : : SNR for a 2-port network : : : : : : : : : : : : Relation between noise gure and SNR : : : : : 4.4 Transmission lines : : : : : : : : : : : : : : : : : : : : : : 4.4.1 Fundamentals of transmission line theory : : : : : : Ideal transmission line : : : : : : : : : : : : : : Non-ideal transmission line : : : : : : : : : : : : Frequency response : : : : : : : : : : : : : : : : Conditions for the absence of signal distortion : : Impulse response of a non-ideal transmission line Secondary constants of some transmission lines : 4.4.2 Cross-talk : : : : : : : : : : : : : : : : : : : : : : Near-end cross-talk : : : : : : : : : : : : : : : : Far-end cross-talk : : : : : : : : : : : : : : : : : 4.5 Optical bers : : : : : : : : : : : : : : : : : : : : : : : : : Description of a ber-optic transmission system : 4.6 Radio links : : : : : : : : : : : : : : : : : : : : : : : : : : 4.6.1 Frequency ranges for radio transmission : : : : : : Radiation masks : : : : : : : : : : : : : : : : : : 4.6.2 Narrowband radio channel model : : : : : : : : : : Equivalent circuit at the receiver : : : : : : : : : Multipath : : : : : : : : : : : : : : : : : : : : : 4.6.3 Doppler shift : : : : : : : : : : : : : : : : : : : : : 4.6.4 Propagation of wideband signals : : : : : : : : : : Channel parameters in the presence of multipath : Statistical description of fading channels : : : : : 4.6.5 Continuous-time channel model : : : : : : : : : : : Power delay prole : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
255 255 255 257 259 259 262 263 263 265 265 265 266 267 268 270 272 272 273 274 275 275 276 279 279 282 282 283 286 288 290 291 292 294 295 296 296 299 299 303 305 307 307 309 310
Contents
xiii
Doppler spectrum : : : : : : : : : : : : : : : : : : : Doppler spectrum models : : : : : : : : : : : : : : : Shadowing : : : : : : : : : : : : : : : : : : : : : : Final remarks : : : : : : : : : : : : : : : : : : : : : 4.6.6 Discrete-time model for fading channels : : : : : : : : Generation of a process with a pre-assigned spectrum 4.7 Telephone channel : : : : : : : : : : : : : : : : : : : : : : : : 4.7.1 Characteristics : : : : : : : : : : : : : : : : : : : : : : Linear distortion : : : : : : : : : : : : : : : : : : : Noise sources : : : : : : : : : : : : : : : : : : : : : Non-linear distortion : : : : : : : : : : : : : : : : : Frequency offset : : : : : : : : : : : : : : : : : : : Phase jitter : : : : : : : : : : : : : : : : : : : : : : Echo : : : : : : : : : : : : : : : : : : : : : : : : : : 4.8 Transmission channel: general model : : : : : : : : : : : : : : Power amplier (HPA) : : : : : : : : : : : : : : : : Transmission medium : : : : : : : : : : : : : : : : : Additive noise : : : : : : : : : : : : : : : : : : : : : Phase noise : : : : : : : : : : : : : : : : : : : : : : Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 5 Digital representation of waveforms 5.1 Analog and digital access : : : : : : : : : : : : : : : : : 5.1.1 Digital representation of speech : : : : : : : : : : Some waveforms : : : : : : : : : : : : : : : : Speech coding : : : : : : : : : : : : : : : : : : The interpolator lter as a holder : : : : : : : : Sizing of the binary channel parameters : : : : 5.1.2 Coding techniques and applications : : : : : : : : 5.2 Instantaneous quantization : : : : : : : : : : : : : : : : : 5.2.1 Parameters of a quantizer : : : : : : : : : : : : : 5.2.2 Uniform quantizers : : : : : : : : : : : : : : : : Quantization error : : : : : : : : : : : : : : : : Relation between 1, b and sat : : : : : : : : Statistical description of the quantization noise Statistical power of the quantization error : : : Design of a uniform quantizer : : : : : : : : : Signal-to-quantization error ratio : : : : : : : : Implementations of uniform PCM encoders : : 5.3 Non-uniform quantizers : : : : : : : : : : : : : : : : : : Three examples of implementation : : : : : : : 5.3.1 Companding techniques : : : : : : : : : : : : : : Signal-to-quantization error ratio : : : : : : : : Digital compression : : : : : : : : : : : : : : : Signal-to-quantization noise ratio mask : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
311 313 313 314 315 316 318 318 319 319 319 319 321 321 322 322 326 326 326 328 331 331 332 332 337 338 340 341 344 344 346 347 350 350 352 353 354 357 358 359 360 364 365 366
xiv
Contents
5.4
5.5
5.6
5.7
5.8
Optimum quantizer in the MSE sense : : : : : : : : : : : : Max algorithm : : : : : : : : : : : : : : : : : : : : : : Lloyd algorithm : : : : : : : : : : : : : : : : : : : : : : Expression of 3q for a very ne quantization : : : : : : Performance of non-uniform quantizers : : : : : : : : : Adaptive quantization : : : : : : : : : : : : : : : : : : : : : : : : General scheme : : : : : : : : : : : : : : : : : : : : : : 5.4.1 Feedforward adaptive quantizer : : : : : : : : : : : : : : : Performance : : : : : : : : : : : : : : : : : : : : : : : : 5.4.2 Feedback adaptive quantizers : : : : : : : : : : : : : : : : Estimate of s .k/ : : : : : : : : : : : : : : : : : : : : : Differential coding (DPCM) : : : : : : : : : : : : : : : : : : : : : 5.5.1 Conguration with feedback quantizer : : : : : : : : : : : 5.5.2 Alternative conguration : : : : : : : : : : : : : : : : : : 5.5.3 Expression of the optimum coefcients : : : : : : : : : : : Effects due to the presence of the quantizer : : : : : : : 5.5.4 Adaptive predictors : : : : : : : : : : : : : : : : : : : : : Adaptive feedforward predictors : : : : : : : : : : : : : Sequential adaptive feedback predictors : : : : : : : : : Performance : : : : : : : : : : : : : : : : : : : : : : : : 5.5.5 Alternative structures for the predictor : : : : : : : : : : : All-pole predictor : : : : : : : : : : : : : : : : : : : : : All-zero predictor : : : : : : : : : : : : : : : : : : : : : Pole-zero predictor : : : : : : : : : : : : : : : : : : : : Pitch predictor : : : : : : : : : : : : : : : : : : : : : : APC : : : : : : : : : : : : : : : : : : : : : : : : : : : : Delta modulation : : : : : : : : : : : : : : : : : : : : : : : : : : 5.6.1 Oversampling and quantization error : : : : : : : : : : : : 5.6.2 Linear delta modulation (LDM) : : : : : : : : : : : : : : : LDM implementation : : : : : : : : : : : : : : : : : : : Choice of system parameters : : : : : : : : : : : : : : : 5.6.3 Adaptive delta modulation (ADM) : : : : : : : : : : : : : Continuously variable slope delta modulation (CVSDM) ADM with second-order predictors : : : : : : : : : : : : 5.6.4 PCM encoder via LDM : : : : : : : : : : : : : : : : : : 5.6.5 Sigma delta modulation (6DM) : : : : : : : : : : : : : : : Coding by modeling : : : : : : : : : : : : : : : : : : : : : : : : : Vocoder or LPC : : : : : : : : : : : : : : : : : : : : : : RPE coding : : : : : : : : : : : : : : : : : : : : : : : : CELP coding : : : : : : : : : : : : : : : : : : : : : : : Multipulse coding : : : : : : : : : : : : : : : : : : : : : Vector quantization (VQ) : : : : : : : : : : : : : : : : : : : : : : 5.8.1 Characterization of VQ : : : : : : : : : : : : : : : : : : : Parameters determining VQ performance : : : : : : : : : Comparison between VQ and scalar quantization : : : : 5.3.2
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
366 369 370 371 374 377 377 379 380 381 382 385 386 389 391 392 393 394 394 398 398 398 399 399 400 401 404 404 407 408 408 410 411 412 412 413 413 414 415 416 417 417 418 418 420
Contents
xv
Optimum quantization : : : : : : : : : : : : : : : Generalized Lloyd algorithm : : : : : : : : : : 5.8.3 LBG algorithm : : : : : : : : : : : : : : : : : : Choice of the initial codebook : : : : : : : : : Description of the LBG algorithm with splitting Selection of the training sequence : : : : : : : 5.8.4 Variants of VQ : : : : : : : : : : : : : : : : : : Tree search VQ : : : : : : : : : : : : : : : : : Multistage VQ : : : : : : : : : : : : : : : : : Product code VQ : : : : : : : : : : : : : : : : 5.9 Other coding techniques : : : : : : : : : : : : : : : : : : Adaptive transform coding (ATC) : : : : : : : Sub-band coding (SBC) : : : : : : : : : : : : : 5.10 Source coding : : : : : : : : : : : : : : : : : : : : : : : 5.11 Speech and audio standards : : : : : : : : : : : : : : : : Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : 5.8.2
: : : : : : : : : : : : : : : : : : : : : : : : procedure : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
421 422 424 425 426 426 429 429 430 430 432 433 433 433 434 435 437 437 439 440 440 442 445 447 449 449 450 452 454 454 455 456 458 459 461 465 470 472 474 475 477 477 478
6 Modulation theory 6.1 Theory of optimum detection : : : : : : : : : : : : : : : : : : Statistics of the random variables fwi g : : : : : : : : Sufcient statistics : : : : : : : : : : : : : : : : : : Decision criterion : : : : : : : : : : : : : : : : : : : Theorem of irrelevance : : : : : : : : : : : : : : : : Implementations of the maximum likelihood criterion Error probability : : : : : : : : : : : : : : : : : : : 6.1.1 Examples of binary signalling : : : : : : : : : : : : : : Antipodal signals ( D 1) : : : : : : : : : : : : : : Orthogonal signals ( D 0) : : : : : : : : : : : : : : Binary FSK : : : : : : : : : : : : : : : : : : : : : : 6.1.2 Limits on the probability of error : : : : : : : : : : : : Upper limit : : : : : : : : : : : : : : : : : : : : : : Lower limit : : : : : : : : : : : : : : : : : : : : : : 6.2 Simplied model of a transmission system and denition of binary channel : : : : : : : : : : : : : : : : : : : : : : : : Parameters of a transmission system : : : : : : : : : Relations among parameters : : : : : : : : : : : : : 6.3 Pulse amplitude modulation (PAM) : : : : : : : : : : : : : : : 6.4 Phase-shift keying (PSK) : : : : : : : : : : : : : : : : : : : : Binary PSK (BPSK) : : : : : : : : : : : : : : : : : Quadrature PSK (QPSK) : : : : : : : : : : : : : : : 6.5 Differential PSK (DPSK) : : : : : : : : : : : : : : : : : : : : 6.5.1 Error probability for an M-DPSK system : : : : : : : : 6.5.2 Differential encoding and coherent demodulation : : : : Binary case (M D 2, differentially encoded BPSK) : Multilevel case : : : : : : : : : : : : : : : : : : : :
xvi
Contents
AM-PM or quadrature amplitude modulation (QAM) : : : : : : : : : : Comparison between PSK and QAM : : : : : : : : : : : : : 6.7 Modulation methods using orthogonal and biorthogonal signals : : : : 6.7.1 Modulation with orthogonal signals : : : : : : : : : : : : : : : Probability of error : : : : : : : : : : : : : : : : : : : : : : Limit of the probability of error for M increasing to innity 6.7.2 Modulation with biorthogonal signals : : : : : : : : : : : : : : Probability of error : : : : : : : : : : : : : : : : : : : : : : 6.8 Binary sequences and coding : : : : : : : : : : : : : : : : : : : : : : Optimum receiver : : : : : : : : : : : : : : : : : : : : : : 6.9 Comparison between coherent modulation methods : : : : : : : : : : : Trade-offs for QAM systems : : : : : : : : : : : : : : : : : Comparison of modulation methods : : : : : : : : : : : : : 6.10 Limits imposed by information theory : : : : : : : : : : : : : : : : : Capacity of a system using amplitude modulation : : : : : : Coding strategies depending on the signal-to-noise ratio : : : Coding gain : : : : : : : : : : : : : : : : : : : : : : : : : : Cut-off rate : : : : : : : : : : : : : : : : : : : : : : : : : : 6.11 Optimum receivers for signals with random phase : : : : : : : : : : : ML criterion : : : : : : : : : : : : : : : : : : : : : : : : : Implementation of a non-coherent ML receiver : : : : : : : Error probability for a non-coherent binary FSK system : : : Performance comparison of binary systems : : : : : : : : : 6.12 Binary modulation systems in the presence of at fading : : : : : : : : Diversity : : : : : : : : : : : : : : : : : : : : : : : : : : : 6.13 Transmission methods : : : : : : : : : : : : : : : : : : : : : : : : : : 6.13.1 Transmission methods between two users : : : : : : : : : : : : Three methods : : : : : : : : : : : : : : : : : : : : : : : : 6.13.2 Channel sharing: deterministic access methods : : : : : : : : : Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : Appendices : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 6.A Gaussian distribution function and Marcum function : : : : : : : : : : 6.A.1 The Q function : : : : : : : : : : : : : : : : : : : : : : : : : 6.A.2 The Marcum function : : : : : : : : : : : : : : : : : : : : : : 6.B Gray coding : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 6.C Baseband PPM and PDM : : : : : : : : : : : : : : : : : : : : : : : : Signal-to-noise ratio : : : : : : : : : : : : : : : : : : : : : 6.D Walsh codes : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 6.6 7 Transmission over dispersive channels 7.1 Baseband digital transmission (PAM systems) : : Transmitter : : : : : : : : : : : : : : : Transmission channel : : : : : : : : : : Receiver : : : : : : : : : : : : : : : : : Power spectral density of a PAM signal : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
480 485 486 486 489 492 493 494 496 498 499 502 502 503 504 506 508 509 509 510 512 516 519 520 521 522 522 523 523 525 527 527 527 529 531 532 532 536 539 539 539 541 542 543
Contents
xvii
7.2
Passband digital transmission (QAM systems) : : : : : : : : : Transmitter : : : : : : : : : : : : : : : : : : : : : : Power spectral density of a QAM signal : : : : : : : Three equivalent representations of the modulator : : Coherent receiver : : : : : : : : : : : : : : : : : : : 7.3 Baseband equivalent model of a QAM system : : : : : : : : : 7.3.1 Signal analysis : : : : : : : : : : : : : : : : : : : : : : Signal-to-noise ratio : : : : : : : : : : : : : : : : : 7.3.2 Characterization of system elements : : : : : : : : : : Transmitter : : : : : : : : : : : : : : : : : : : : : : Transmission channel : : : : : : : : : : : : : : : : : Receiver : : : : : : : : : : : : : : : : : : : : : : : : 7.3.3 Intersymbol interference : : : : : : : : : : : : : : : : : Discrete-time equivalent system : : : : : : : : : : : Nyquist pulses : : : : : : : : : : : : : : : : : : : : Eye diagram : : : : : : : : : : : : : : : : : : : : : : 7.3.4 Performance analysis : : : : : : : : : : : : : : : : : : Symbol error probability in the absence of ISI : : : : Matched lter receiver : : : : : : : : : : : : : : : : 7.4 Carrierless AM/PM (CAP) modulation : : : : : : : : : : : : : 7.5 Regenerative PCM repeaters : : : : : : : : : : : : : : : : : : : 7.5.1 PCM signals over a binary channel : : : : : : : : : : : Linear PCM coding of waveforms : : : : : : : : : : Overall system performance : : : : : : : : : : : : : 7.5.2 Regenerative repeaters : : : : : : : : : : : : : : : : : : Analog transmission : : : : : : : : : : : : : : : : : Digital transmission : : : : : : : : : : : : : : : : : : Comparison between analog and digital transmission Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : Appendices : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 7.A Line codes for PAM systems : : : : : : : : : : : : : : : : : : 7.A.1 Line codes : : : : : : : : : : : : : : : : : : : : : : : : Non-return-to-zero (NRZ) format : : : : : : : : : : : Return-to-zero (RZ) format : : : : : : : : : : : : : : Biphase (B- ) format : : : : : : : : : : : : : : : : : Delay modulation or Miller code : : : : : : : : : : : Block line codes : : : : : : : : : : : : : : : : : : : Alternate mark inversion (AMI) : : : : : : : : : : : 7.A.2 Partial response systems : : : : : : : : : : : : : : : : : The choice of the PR polynomial : : : : : : : : : : : Symbol detection and error probability : : : : : : : : Precoding : : : : : : : : : : : : : : : : : : : : : : : Error probability with precoding : : : : : : : : : : : Alternative interpretation of PR systems : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
544 544 546 547 548 549 550 552 553 553 553 555 556 556 559 562 565 565 567 568 571 571 572 573 575 576 577 578 581 583 583 583 583 584 584 585 585 586 587 590 594 596 597 599
xviii
Contents
7.B Computation of Pe for some cases of interest 7.B.1 Pe in the absence of ISI : : : : : : : 7.B.2 Pe in the presence of ISI : : : : : : Exhaustive method : : : : : : : : Gaussian approximation : : : : : Worst-case limit : : : : : : : : : : Saltzberg limit : : : : : : : : : : GQR method : : : : : : : : : : : 7.C Coherent PAM-DSB transmission : : : : : : General scheme : : : : : : : : : : Transmit signal PSD : : : : : : : Signal-to-noise ratio : : : : : : : 7.D Implementation of a QAM transmitter : : : 7.E Simulation of a QAM system : : : : : : : :
: : : : : : : : : : : : : :
: : : : : : : : : : : : : :
: : : : : : : : : : : : : :
: : : : : : : : : : : : : :
: : : : : : : : : : : : : :
: : : : : : : : : : : : : :
: : : : : : : : : : : : : :
: : : : : : : : : : : : : :
: : : : : : : : : : : : : :
: : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
602 602 604 604 605 605 606 607 608 608 609 609 611 613 619 619 620 620 622 626 627 628 630 633 635 638 639 642 644 645 645 646 648 648 649 649 655 655 657 657 658 659 662 663
8 Channel equalization and symbol detection 8.1 Zero-forcing equalizer (LE-ZF) : : : : : : : : : : : : : : : : : 8.2 Linear equalizer (LE) : : : : : : : : : : : : : : : : : : : : : : 8.2.1 Optimum receiver in the presence of noise and ISI : : : Alternative derivation of the IIR equalizer : : : : : : Signal-to-noise ratio : : : : : : : : : : : : : : : : 8.3 LE with a nite number of coefcients : : : : : : : : : : : : : Adaptive LE : : : : : : : : : : : : : : : : : : : : : 8.4 Fractionally spaced equalizer (FSE) : : : : : : : : : : : : : : : Adaptive FSE : : : : : : : : : : : : : : : : : : : : : 8.5 Decision feedback equalizer (DFE) : : : : : : : : : : : : : : : Adaptive DFE : : : : : : : : : : : : : : : : : : : : : Design of a DFE with a nite number of coefcients Design of a fractionally spaced DFE (FS-DFE) : : : Signal-to-noise ratio : : : : : : : : : : : : : : : : Remarks : : : : : : : : : : : : : : : : : : : : : : : : 8.6 Convergence behavior of adaptive equalizers : : : : : : : : : : Adaptive LE : : : : : : : : : : : : : : : : : : : : : Adaptive DFE : : : : : : : : : : : : : : : : : : : : : 8.7 LE-ZF with a nite number of coefcients : : : : : : : : : : : 8.8 DFE: alternative congurations : : : : : : : : : : : : : : : : : DFE-ZF : : : : : : : : : : : : : : : : : : : : : : : : DFE-ZF as a noise predictor : : : : : : : : : : : : : DFE as ISI and noise predictor : : : : : : : : : : : : 8.9 Benchmark performance for two equalizers : : : : : : : : : : : Performance comparison : : : : : : : : : : : : : : : Equalizer performance for two channel models : : : 8.10 Optimum methods for data detection : : : : : : : : : : : : : : 8.10.1 Maximum likelihood sequence detection : : : : : : : : Lower limit to error probability using the MLSD criterion : : : : : : : : : : :
: : : : :
Contents
xix
The Viterbi algorithm (VA) : : : : : : : : : : : : : : : : : : : Computational complexity of the VA : : : : : : : : : : : : : : 8.10.2 Maximum a posteriori probability detector : : : : : : : : : : : : Statistical description of a sequential machine : : : : : : : : : The forward-backward algorithm (FBA) : : : : : : : : : : : : Scaling : : : : : : : : : : : : : : : : : : : : : : : : : : : : : Likelihood function in the absence of ISI : : : : : : : : : : : Simplied version of the MAP algorithm (Max-Log-MAP) : : Relation between Max-Log-MAP and Log-MAP : : : : : : : : 8.11 Optimum receivers for transmission over dispersive channels : : : : : : Ungerboecks formulation of the MLSD : : : : : : : : : : : : 8.12 Error probability achieved by MLSD : : : : : : : : : : : : : : : : : : : Computation of the minimum distance : : : : : : : : : : : : : 8.13 Reduced state sequence detection : : : : : : : : : : : : : : : : : : : : : Reduced state trellis diagram : : : : : : : : : : : : : : : : : : RSSE algorithm : : : : : : : : : : : : : : : : : : : : : : : : : Further simplication: DFSE : : : : : : : : : : : : : : : : : : 8.14 Passband equalizers : : : : : : : : : : : : : : : : : : : : : : : : : : : : 8.14.1 Passband receiver structure : : : : : : : : : : : : : : : : : : : : Joint optimization of equalizer coefcients and carrier phase offset : : : : : : : : : : : : : : Adaptive method : : : : : : : : : : : : : : : : : : : : : : : : 8.14.2 Efcient implementations of voiceband modems : : : : : : : : : 8.15 LE for voiceband modems : : : : : : : : : : : : : : : : : : : : : : : : : Detection of the training sequence : : : : : : : : : : : : : : : Computations of the coefcients of a cyclic equalizer : : : : : Transition from training to data mode : : : : : : : : : : : : : Example of application: a simple modem : : : : : : : : : : : 8.16 LE and DFE in the frequency domain with data frames using cyclic prex 8.17 Numerical results obtained by simulations : : : : : : : : : : : : : : : : QPSK transmission over a minimum phase channel : : : : : : QPSK transmission over a non-minimum phase channel : : : : 8-PSK transmission over a minimum phase channel : : : : : : 8-PSK transmission over a non-minimum phase channel : : : 8.18 Diversity combining techniques : : : : : : : : : : : : : : : : : : : : : : Antenna arrays : : : : : : : : : : : : : : : : : : : : : : : : : Combining techniques : : : : : : : : : : : : : : : : : : : : : Equalization and diversity : : : : : : : : : : : : : : : : : : : Diversity in transmission : : : : : : : : : : : : : : : : : : : : Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : Appendices : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 8.A Calculus of variations and receiver optimization : : : : : : : : : : : : : 8.A.1 Calculus of variations : : : : : : : : : : : : : : : : : : : : : : : Linear functional : : : : : : : : : : : : : : : : : : : : : : : : Quadratic functional : : : : : : : : : : : : : : : : : : : : : :
663 667 668 668 670 673 674 675 677 678 680 682 686 691 691 694 695 697 698 700 701 703 705 706 707 709 709 710 713 713 715 716 716 717 718 719 722 722 726 731 731 731 731 732
xx
Contents
8.A.2 Receiver optimization : : : : : : : : : : : : : : 8.A.3 Joint optimization of transmitter and receiver : : 8.B DFE design: matrix formulations : : : : : : : : : : : : 8.B.1 Method based on correlation sequences : : : : : 8.B.2 Method based on the channel impulse response and i.i.d. symbols : : : : : : : : : : : : : : : : 8.B.3 Method based on the channel impulse response and any symbol statistic : : : : : : : : : : : : : 8.B.4 FS-DFE : : : : : : : : : : : : : : : : : : : : : 8.C Equalization based on the peak value of ISI : : : : : : 8.D Description of a nite state machine (FSM) : : : : : : :
: : : :
: : : :
: : : :
: : : :
: : : :
: : : :
: : : :
: : : :
: : : :
735 739 741 741 744 746 747 749 751 753 753 755 755 755 755 756 757 759 760 764 769 770 771 771 773 773 775 777 779 780 780 781 781 782 783 786 788 793 795 795 795
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
9 Orthogonal frequency division multiplexing 9.1 OFDM systems : : : : : : : : : : : : : : : : : : : : : : : : 9.2 Orthogonality conditions : : : : : : : : : : : : : : : : : : : : Time domain : : : : : : : : : : : : : : : : : : : : Frequency domain : : : : : : : : : : : : : : : : : z-transform domain : : : : : : : : : : : : : : : : : 9.3 Efcient implementation of OFDM systems : : : : : : : : : OFDM implementation employing matched lters : Orthogonality conditions in terms of the polyphase components : : : : : OFDM implementation employing a prototype lter 9.4 Non-critically sampled lter banks : : : : : : : : : : : : : : 9.5 Examples of OFDM systems : : : : : : : : : : : : : : : : : Discrete multitone (DMT) : : : : : : : : : : : : : Filtered multitone (FMT) : : : : : : : : : : : : : : Discrete wavelet multitone (DWMT) : : : : : : : : 9.6 Equalization of OFDM systems : : : : : : : : : : : : : : : : Interpolator lter and virtual subchannels : : : : : Equalization of DMT systems : : : : : : : : : : : Equalization of FMT systems : : : : : : : : : : : : 9.7 Synchronization of OFDM systems : : : : : : : : : : : : : : 9.8 Passband OFDM systems : : : : : : : : : : : : : : : : : : : Passband DWMT systems : : : : : : : : : : : : : Passband DMT and FMT systems : : : : : : : : : Comparison between OFDM and QAM systems : : 9.9 DWMT modulation : : : : : : : : : : : : : : : : : : : : : : Transmit and receive lter banks : : : : : : : : : : Approximate interchannel interference suppression Perfect interchannel interference suppression : : : : Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
10 Spread spectrum systems 10.1 Spread spectrum techniques : : : : : : : : : : : : : : : : : : : : : : : : 10.1.1 Direct sequence systems : : : : : : : : : : : : : : : : : : : : : :
Contents
xxi
Classication of CDMA systems : : : : : : : : : Synchronization : : : : : : : : : : : : : : : : : : 10.1.2 Frequency hopping systems : : : : : : : : : : : : : Classication of FH systems : : : : : : : : : : : 10.2 Applications of spread spectrum systems : : : : : : : : : : 10.2.1 Anti-jam communications : : : : : : : : : : : : : : 10.2.2 Multiple-access systems : : : : : : : : : : : : : : : 10.2.3 Interference rejection : : : : : : : : : : : : : : : : 10.3 Chip matched lter and rake receiver : : : : : : : : : : : : Number of resolvable rays in a multipath channel Chip matched lter (CMF) : : : : : : : : : : : : 10.4 Interference : : : : : : : : : : : : : : : : : : : : : : : : : Detection strategies for multiple-access systems : 10.5 Equalizers for single-user detection : : : : : : : : : : : : : Chip equalizer (CE) : : : : : : : : : : : : : : : : Symbol equalizer (SE) : : : : : : : : : : : : : : 10.6 Block equalizer for multiuser detection : : : : : : : : : : : 10.7 Maximum likelihood multiuser detector : : : : : : : : : : : Correlation matrix approach : : : : : : : : : : : Whitening lter approach : : : : : : : : : : : : : Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
802 804 804 806 807 808 810 811 811 811 813 816 818 818 818 819 820 823 823 824 824 827 828 830 830 830 833 836 837 837 838 839 842 843 845 846 849 851 854 857 859 861 862 862
11 Channel codes 11.1 System model : : : : : : : : : : : : : : : : : : : : : : : : : : : : 11.2 Block codes : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 11.2.1 Theory of binary codes with group structure : : : : : : : : Properties : : : : : : : : : : : : : : : : : : : : : : : : : Parity check matrix : : : : : : : : : : : : : : : : : : : : Code generator matrix : : : : : : : : : : : : : : : : : : Decoding of binary parity check codes : : : : : : : : : : Cosets : : : : : : : : : : : : : : : : : : : : : : : : : : : Two conceptually simple decoding methods : : : : : : : Syndrome decoding : : : : : : : : : : : : : : : : : : : : 11.2.2 Fundamentals of algebra : : : : : : : : : : : : : : : : : : Modulo q arithmetic : : : : : : : : : : : : : : : : : : : Polynomials with coefcients from a eld : : : : : : : : The concept of modulo in the arithmetic of polynomials Devices to sum and multiply elements in a nite eld : : Remarks on nite elds : : : : : : : : : : : : : : : : : : Roots of a polynomial : : : : : : : : : : : : : : : : : : Minimum function : : : : : : : : : : : : : : : : : : : : Methods to determine the minimum function : : : : : : Properties of the minimum function : : : : : : : : : : : 11.2.3 Cyclic codes : : : : : : : : : : : : : : : : : : : : : : : : : The algebra of cyclic codes : : : : : : : : : : : : : : : :
xxii
Contents
Properties of cyclic codes : : : : : : : : : : : : : : : : Encoding method using a shift register of length r : : Encoding method using a shift register of length k : : Hard decoding of cyclic codes : : : : : : : : : : : : : Hamming codes : : : : : : : : : : : : : : : : : : : : : Burst error detection : : : : : : : : : : : : : : : : : : 11.2.4 Simplex cyclic codes : : : : : : : : : : : : : : : : : : : Relation to PN sequences : : : : : : : : : : : : : : : : 11.2.5 BCH codes : : : : : : : : : : : : : : : : : : : : : : : : An alternative method to specify the code polynomials BoseChaudhuriHocquenhem (BCH) codes : : : : : : Binary BCH codes : : : : : : : : : : : : : : : : : : : ReedSolomon codes : : : : : : : : : : : : : : : : : : Decoding of BCH codes : : : : : : : : : : : : : : : : Efcient decoding of BCH codes : : : : : : : : : : : : 11.2.6 Performance of block codes : : : : : : : : : : : : : : : : 11.3 Convolutional codes : : : : : : : : : : : : : : : : : : : : : : : : 11.3.1 General description of convolutional codes : : : : : : : : Parity check matrix : : : : : : : : : : : : : : : : : : : Generator matrix : : : : : : : : : : : : : : : : : : : : Transfer function : : : : : : : : : : : : : : : : : : : : Catastrophic error propagation : : : : : : : : : : : : : 11.3.2 Decoding of convolutional codes : : : : : : : : : : : : : Interleaving : : : : : : : : : : : : : : : : : : : : : : : Two decoding models : : : : : : : : : : : : : : : : : : Viterbi algorithm : : : : : : : : : : : : : : : : : : : : Forward-backward algorithm : : : : : : : : : : : : : : Sequential decoding : : : : : : : : : : : : : : : : : : : 11.3.3 Performance of convolutional codes : : : : : : : : : : : 11.4 Concatenated codes : : : : : : : : : : : : : : : : : : : : : : : : Soft-output Viterbi algorithm (SOVA) : : : : : : : : : 11.5 Turbo codes : : : : : : : : : : : : : : : : : : : : : : : : : : : : Encoding : : : : : : : : : : : : : : : : : : : : : : : : The basic principle of iterative decoding : : : : : : : : The forward-backward algorithm revisited : : : : : : : Iterative decoding : : : : : : : : : : : : : : : : : : : : Performance evaluation : : : : : : : : : : : : : : : : : 11.6 Iterative detection and decoding : : : : : : : : : : : : : : : : : 11.7 Low-density parity check codes : : : : : : : : : : : : : : : : : : Encoding procedure : : : : : : : : : : : : : : : : : : : Decoding algorithm : : : : : : : : : : : : : : : : : : : Example of application : : : : : : : : : : : : : : : : : Performance and coding gain : : : : : : : : : : : : : : Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : Appendices : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
864 869 870 871 872 875 875 877 878 878 880 883 885 887 891 899 900 903 905 906 907 910 912 913 913 915 915 917 919 921 921 924 924 929 930 939 941 943 946 948 948 953 954 956 960
Contents
xxiii
11.A Nonbinary parity check codes : : : : : : : : : : : : : Linear codes : : : : : : : : : : : : : : : : Parity check matrix : : : : : : : : : : : : : Code generator matrix : : : : : : : : : : : Decoding of nonbinary parity check codes : Coset : : : : : : : : : : : : : : : : : : : : Two conceptually simple decoding methods Syndrome decoding : : : : : : : : : : : : :
: : : : : : : :
: : : : : : : :
: : : : : : : :
: : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
960 961 962 963 964 964 965 965 967 968 968 970 970 973 975 980 985 987 990 990 993 995 996 999 999 1002 1002 1003 1007 1008 1009 1012 1018 1025 1027 1027 1029 1031 1032 1034 1035 1036 1040
12 Trellis coded modulation 12.1 Linear TCM for one- and two-dimensional signal sets : : : : 12.1.1 Fundamental elements : : : : : : : : : : : : : : : : : Basic TCM scheme : : : : : : : : : : : : : : : : : Example : : : : : : : : : : : : : : : : : : : : : : : 12.1.2 Set partitioning : : : : : : : : : : : : : : : : : : : : 12.1.3 Lattices : : : : : : : : : : : : : : : : : : : : : : : : 12.1.4 Assignment of symbols to the transitions in the trellis 12.1.5 General structure of the encoder/bit-mapper : : : : : Computation of dfree : : : : : : : : : : : : : : : : 12.2 Multidimensional TCM : : : : : : : : : : : : : : : : : : : : Encoding : : : : : : : : : : : : : : : : : : : : : : Decoding : : : : : : : : : : : : : : : : : : : : : : 12.3 Rotationally invariant TCM schemes : : : : : : : : : : : : : Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 13 Precoding and coding techniques for dispersive channels 13.1 Capacity of a dispersive channel : : : : : : : : : : : : : 13.2 Techniques to achieve capacity : : : : : : : : : : : : : : Bit loading for OFDM : : : : : : : : : : : : : Discrete-time model of a single carrier system : Achieving capacity with a single carrier system 13.3 Precoding and coding for dispersive channels : : : : : : : 13.3.1 TomlinsonHarashima (TH) precoding : : : : : : 13.3.2 TH precoding and TCM : : : : : : : : : : : : : : 13.3.3 Flexible precoding : : : : : : : : : : : : : : : : : Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
14 Synchronization 14.1 The problem of synchronization for QAM systems : : : : : : 14.2 The phase-locked loop : : : : : : : : : : : : : : : : : : : : : 14.2.1 PLL baseband model : : : : : : : : : : : : : : : : : Linear approximation : : : : : : : : : : : : : : : : 14.2.2 Analysis of the PLL in the presence of additive noise Noise analysis using the linearity assumption : : : 14.2.3 Analysis of a second-order PLL : : : : : : : : : : : : 14.3 Costas loop : : : : : : : : : : : : : : : : : : : : : : : : : :
xxiv
Contents
14.3.1 PAM signals : : : : : : : : : : : : : : : : : : : : : : : : 14.3.2 QAM signals : : : : : : : : : : : : : : : : : : : : : : : 14.4 The optimum receiver : : : : : : : : : : : : : : : : : : : : : : : Timing recovery : : : : : : : : : : : : : : : : : : : : Carrier phase recovery : : : : : : : : : : : : : : : : : 14.5 Algorithms for timing and carrier phase recovery : : : : : : : : : 14.5.1 ML criterion : : : : : : : : : : : : : : : : : : : : : : : : Assumption of slow time varying channel : : : : : : : 14.5.2 Taxonomy of algorithms using the ML criterion : : : : : Feedback estimators : : : : : : : : : : : : : : : : : : Early-late estimators : : : : : : : : : : : : : : : : : : 14.5.3 Timing estimators : : : : : : : : : : : : : : : : : : : : : Non-data aided : : : : : : : : : : : : : : : : : : : : : Non-data aided via spectral estimation : : : : : : : : : Data-aided and data-directed : : : : : : : : : : : : : : Data- and phase-directed with feedback: differentiator scheme : : : : : : : : : : : : Data- and phase-directed with feedback: Mueller & Muller scheme : : : : : : : : : Non-data aided with feedback : : : : : : : : : : : : : 14.5.4 Phasor estimators : : : : : : : : : : : : : : : : : : : : : Data- and timing-directed : : : : : : : : : : : : : : : : Non-data aided for M-PSK signals : : : : : : : : : : : Data- and timing-directed with feedback : : : : : : : : 14.6 Algorithms for carrier frequency recovery : : : : : : : : : : : : 14.6.1 Frequency offset estimators : : : : : : : : : : : : : : : : Non-data aided : : : : : : : : : : : : : : : : : : : : : Non-data aided and timing-independent with feedback : Non-data aided and timing-directed with feedback : : : 14.6.2 Estimators operating at the modulation rate : : : : : : : : Data-aided and data-directed : : : : : : : : : : : : : : Non-data aided for M-PSK : : : : : : : : : : : : : : : 14.7 Second-order digital PLL : : : : : : : : : : : : : : : : : : : : : 14.8 Synchronization in spread spectrum systems : : : : : : : : : : : 14.8.1 The transmission system : : : : : : : : : : : : : : : : : Transmitter : : : : : : : : : : : : : : : : : : : : : : : Optimum receiver : : : : : : : : : : : : : : : : : : : : 14.8.2 Timing estimators with feedback : : : : : : : : : : : : : Non-data aided: non-coherent DLL : : : : : : : : : : : Non-data aided MCTL : : : : : : : : : : : : : : : : : Data- and phase-directed: coherent DLL : : : : : : : : Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : :
: : : : : : : : : : : : : : :
: : : : : : : : : : : : : : :
: : : : : : : : : : : : : : :
1040 1042 1044 1046 1050 1051 1051 1051 1051 1053 1055 1055 1055 1057 1059 1062 1064 1065 1066 1066 1066 1067 1068 1069 1069 1071 1071 1072 1073 1073 1074 1074 1074 1074 1075 1076 1076 1077 1077 1081 1083 1083 1086
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
15 Self-training equalization 15.1 Problem denition and fundamentals : : : : : : : : : : : : : : : : : : : Minimization of a special function : : : : : : : : : : : : : : :
Contents
xxv
15.2 Three algorithms for PAM systems : : : : : : : : : : : : : The Sato algorithm : : : : : : : : : : : : : : : : BenvenisteGoursat algorithm : : : : : : : : : : Stop-and-go algorithm : : : : : : : : : : : : : : Remarks : : : : : : : : : : : : : : : : : : : : : : 15.3 The contour algorithm for PAM systems : : : : : : : : : : Simplied realization of the contour algorithm : : 15.4 Self-training equalization for partial response systems : : : The Sato algorithm for partial response systems : Contour algorithm for partial response systems : 15.5 Self-training equalization for QAM systems : : : : : : : : The Sato algorithm for QAM systems : : : : : : 15.5.1 Constant modulus algorithm : : : : : : : : : : : : : The contour algorithm for QAM systems : : : : Joint contour algorithm and carrier phase tracking 15.6 Examples of applications : : : : : : : : : : : : : : : : : : Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : : Appendices : : : : : : : : : : : : : : : : : : : : : : : : : : : : 15.A On the convergence of the contour algorithm : : : : : : : :
: : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
1090 1090 1091 1092 1092 1093 1095 1096 1096 1098 1100 1100 1101 1102 1104 1106 1111 1113 1113 1115 1116 1117 1118 1119 1120 1121 1124 1128 1131 1136 1137 1142 1145 1145 1145 1145 1146 1152 1156 1158
16 Applications of interference cancellation 16.1 Echo and nearend cross-talk cancellation for PAM systems : Cross-talk cancellation and full duplex transmission Polyphase structure of the canceller : : : : : : : : Canceller at symbol rate : : : : : : : : : : : : : : Adaptive canceller : : : : : : : : : : : : : : : : : Canceller structure with distributed arithmetic : : : 16.2 Echo cancellation for QAM systems : : : : : : : : : : : : : 16.3 Echo cancellation for OFDM systems : : : : : : : : : : : : : 16.4 Multiuser detection for VDSL : : : : : : : : : : : : : : : : : 16.4.1 Upstream power back-off : : : : : : : : : : : : : : : 16.4.2 Comparison of PBO methods : : : : : : : : : : : : : Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
17 Wired and wireless network technologies 17.1 Wired network technologies : : : : : : : : : : : : : : : : : : : : : 17.1.1 Transmission over unshielded twisted pairs in the customer service area : : : : : : : : : : : : : : : : : : : : : : : : : Modem : : : : : : : : : : : : : : : : : : : : : : : : : : Digital subscriber line : : : : : : : : : : : : : : : : : : 17.1.2 High speed transmission over unshielded twisted pairs in local area networks : : : : : : : : : : : : : : : : : : : : 17.1.3 Hybrid ber/coaxial cable networks : : : : : : : : : : : : : Ranging and power adjustment for uplink transmission : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
xxvi
Contents
17.2 Wireless network technologies : : : : : : : 17.2.1 Wireless local area networks : : : Medium access control protocols 17.2.2 MMDS and LMDS : : : : : : : : Bibliography : : : : : : : : : : : : : : : : : : : Appendices : : : : : : : : : : : : : : : : : : : 17.A Standards for wireless systems : : : : : : 17.A.1 General observations : : : : : : : Wireless systems : : : : : : : : Modulation techniques : : : : : Parameters of the modulator : : Cells in a wireless system : : : 17.A.2 GSM standard : : : : : : : : : : : System characteristics : : : : : : Radio subsystem : : : : : : : : GSM-EDGE : : : : : : : : : : : 17.A.3 IS-136 standard : : : : : : : : : : 17.A.4 JDC standard : : : : : : : : : : : 17.A.5 IS-95 standard : : : : : : : : : : : 17.A.6 DECT standard : : : : : : : : : : 17.A.7 HIPERLAN standard : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : :
1161 1162 1164 1165 1167 1170 1170 1171 1171 1171 1171 1172 1172 1172 1175 1177 1177 1180 1180 1182 1185 1189 1189 1189 1190 1190 1191 1192 1192 1192 1194 1196 1197 1198 1198 1198 1200 1201 1202 1203 1203 1203 1207 1207
18 Modulation techniques for wireless systems 18.1 Analog front-end architectures : : : : : : : : : : : : : : : : : Conventional superheterodyne receiver : : : : : : : Alternative architectures : : : : : : : : : : : : : : Direct conversion receiver : : : : : : : : : : : : : Single conversion to low-IF : : : : : : : : : : : : Double conversion and wideband IF : : : : : : : : 18.2 Three non-coherent receivers for phase modulation systems : 18.2.1 Baseband differential detector : : : : : : : : : : : : : 18.2.2 IF-band (1 Bit) differential detector (1BDD) : : : : : Performance of M-DPSK : : : : : : : : : : : : : : 18.2.3 FM discriminator with integrate and dump lter (LDI) 18.3 Variants of QPSK : : : : : : : : : : : : : : : : : : : : : : : 18.3.1 Basic schemes : : : : : : : : : : : : : : : : : : : : : QPSK : : : : : : : : : : : : : : : : : : : : : : : : Offset QPSK or staggered QPSK : : : : : : : : : : Differential QPSK (DQPSK) : : : : : : : : : : : : =4-DQPSK : : : : : : : : : : : : : : : : : : : : 18.3.2 Implementations : : : : : : : : : : : : : : : : : : : : QPSK, OQPSK, and DQPSK modulators : : : : : =4-DQPSK modulators : : : : : : : : : : : : : : 18.4 Frequency shift keying (FSK) : : : : : : : : : : : : : : : : : 18.4.1 Power spectrum of M-FSK : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : :
Contents
xxvii
Power spectrum of non-coherent binary FSK : : Power spectrum of coherent M-FSK : : : : : : : 18.4.2 FSK receivers and corresponding performance : : : Coherent demodulator : : : : : : : : : : : : : : Non-coherent demodulator : : : : : : : : : : : : Limiter-discriminator FM demodulator : : : : : : 18.5 Minimum shift keying (MSK) : : : : : : : : : : : : : : : : 18.5.1 Power spectrum of continuous-phase FSK (CPFSK) 18.5.2 The MSK signal viewed from two perspectives : : Phase of an MSK signal : : : : : : : : : : : : : MSK as binary CPFSK : : : : : : : : : : : : : : MSK as OQPSK : : : : : : : : : : : : : : : : : Complex notation of an MSK signal : : : : : : : 18.5.3 Implementations of an MSK scheme : : : : : : : : 18.5.4 Performance of MSK demodulators : : : : : : : : : MSK with differential precoding : : : : : : : : : 18.5.5 Remarks on spectral containment : : : : : : : : : : 18.6 Gaussian MSK (GMSK) : : : : : : : : : : : : : : : : : : 18.6.1 GMSK via CPFSK : : : : : : : : : : : : : : : : : 18.6.2 Power spectrum of GMSK : : : : : : : : : : : : : 18.6.3 Implementation of a GMSK scheme : : : : : : : : Conguration I : : : : : : : : : : : : : : : : : : Conguration II : : : : : : : : : : : : : : : : : : Conguration III : : : : : : : : : : : : : : : : : 18.6.4 Linear approximation of a GMSK signal : : : : : : Performance of GMSK demodulators : : : : : : Performance of a GSM receiver in the presence of multipath : : : : : : : : Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : : Appendices : : : : : : : : : : : : : : : : : : : : : : : : : : : : 18.A Continuous phase modulation (CPM) : : : : : : : : : : : : Alternative denition of CPM : : : : : : : : : : Advantages of CPM : : : : : : : : : : : : : : : 19 Design of high speed transmission systems over unshielded twisted pair cables 19.1 Design of a quaternary partial response class-IV system sion at 125 Mbit/s : : : : : : : : : : : : : : : : : : : : Analog lter design : : : : : : : : : : : : : : Received signal and adaptive gain control : : Near-end cross-talk cancellation : : : : : : : Decorrelation lter : : : : : : : : : : : : : : Adaptive equalizer : : : : : : : : : : : : : : Compensation of the timing phase drift : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :
1208 1209 1212 1212 1213 1213 1214 1217 1217 1217 1219 1220 1222 1224 1224 1227 1228 1229 1229 1231 1234 1234 1234 1236 1238 1238 1243 1244 1246 1246 1246 1248
1249 for data transmis: : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 1249 1249 1250 1251 1251 1252 1252
xxviii
Contents
Adaptive equalizer coefcient adaptation : : : : : : Convergence behavior of the various algorithms : : 19.1.1 VLSI implementation : : : : : : : : : : : : : : : : : Adaptive digital NEXT canceller : : : : : : : : : : Adaptive digital equalizer : : : : : : : : : : : : : : Timing control : : : : : : : : : : : : : : : : : : : Viterbi detector : : : : : : : : : : : : : : : : : : : 19.2 Design of a dual duplex transmission system at 100 Mbit/s : : Dual duplex transmission : : : : : : : : : : : : : : Physical layer control : : : : : : : : : : : : : : : : Coding and decoding : : : : : : : : : : : : : : : : 19.2.1 Signal processing functions : : : : : : : : : : : : : : The 100BASE-T2 transmitter : : : : : : : : : : : : The 100BASE-T2 receiver : : : : : : : : : : : : : Computational complexity of digital receive lters : Bibliography : : : : : : : : : : : : : : : : : : : : : : : : : : : : : Appendices : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 19.A Interference suppression : : : : : : : : : : : : : : : : : : : : Index
: : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : :
: : : : : : : : : : : : : : : : : :
1253 1253 1255 1255 1258 1261 1262 1263 1263 1265 1266 1269 1269 1270 1272 1273 1274 1274 1277
Preface
The motivation for this book is twofold. On the one hand, we provide a didactic tool to students of communications systems. On the other hand, we present a discussion of fundamental algorithms and structures for telecommunication technologies. The contents reect our experience in teaching courses on Algorithms for Telecommunications at the University of Padova, Italy, as well as our professional experience acquired in industrial research laboratories. The text explains the procedures for solving problems posed by the design of systems for reliable communications over wired or wireless channels. In particular, we focus on fundamental developments in the eld in order to provide the reader with the necessary insight to design essential elements of various communications systems. The book is divided into nineteen chapters. We briey indicate four tracks corresponding to specic areas and course work offered. Track 1. Track 1 includes the basic elements for a rst course on telecommunications, which we regard as an introduction to the remaining tracks. It covers Chapter 1, which recalls fundamental concepts on signals and random processes, with an emphasis on secondorder statistical descriptions. A discussion of the characteristics of transmission media follows in Chapter 4. In this track we focus on the description of noise in electronic devices and on the laws of propagation in transmission lines and radio channels. The representation of waveforms by sequences of binary symbols is treated in Chapter 5; for a rst course it is suggested that emphasis be placed on PCM. Next, Chapter 6 examines the fundamental principles of a digital transmission system, where a sequence of information symbols is sent over a transmission channel. We refer to Shannon theorem to establish the maximum bit rate that can be transmitted reliably over a noisy channel. Signal dispersion caused by a transmission channel is then analyzed in Chapter 7. Examples of elementary and practical implementations of transmission systems are presented, together with a brief introduction to computer simulations. The rst three sections of Chapter 11, where we introduce methods for increasing transmission reliability by exploiting the redundancy added to the information bits, conclude the rst track. Track 2. Track 2, which is an extension of Track 1, focuses on modulation techniques. First, parametric models of random processes are analyzed in Chapter 1. The Wiener lter and the linear prediction theory, which constitute fundamental elements for receiver design, are dealt with in Chapter 2. Chapter 3 lists iterative methods to achieve the objectives stated
xxx
Preface
in Chapter 2, as well as various applications of the Wiener lter, for example channel identication and interference cancellation. These applications are further developed in the rst two sections of Chapter 16. In the rst part of Chapter 8, channel equalization is examined as a further application of the Wiener lter. In the second part of the chapter, more sophisticated methods of equalization and symbol detection, which rely on the Viterbi algorithm and on the forwardbackward algorithm, are analyzed. Initially single-carrier modulation systems are considered. In Chapter 9, we introduce multicarrier modulation techniques, which are preferable for transmission over very dispersive channels and/or applications that require exibility in spectral allocation. In Chapter 10 spread spectrum systems are examined, with emphasis to applications for simultaneous channel access by several users that share a wideband channel. The inherent narrowband interference rejection capabilities of spread spectrum systems, as well as their implementations, are also discussed. This is followed by Chapter 18, which illustrates specic modulation techniques developed for mobile radio applications. Track 3. We observe the trend towards implementing transceiver functions using digital signal processors. Therefore the algorithmic aspects of a transmission system are becoming increasingly important. Hardware devices are assigned wherever possible only the functions of analog front-end, xed ltering, and digital-to-analog and analog-to-digital conversion. This approach enhances the exibility of transceivers, which can be utilized for more than one transmission standard, and considerably reduces development time. In line with the above considerations, Track 3 begins with a review of Chapters 2 and 3, which illustrate the fundamental principles of transmission system design, and of Chapter 8, which investigates individual building blocks for channel equalization and symbol detection. The assumption that the transmission channel characteristics are known a priori is removed in Chapter 15, where blind equalization techniques are discussed. Channel coding techniques to improve the reliability of transmission are investigated in depth in Chapters 11 and 12. A further method to mitigate channel dispersion is precoding. The operations of systems that employ joint precoding and channel coding are explained in Chapter 13. Because of electromagnetic coupling, the desired signal at the receiver is often disturbed by other transmissions taking place simultaneously. Cancellation techniques to suppress interference signals are treated in Chapter 16. Track 4. Track 4 addresses various challenges encountered in designing wired and wireless communications systems. The elements introduced in Chapters 2 and 3, as well as the algorithms introduced in Chapter 8, are essential for this track. The principles of multicarrier and spread spectrum modulation techniques, which are increasingly being adopted in communications systems, are investigated in depth in Chapters 9 and 10, respectively. The design of the receiver front-end, as well as various methods for timing and carrier recovery, are dealt with in Chapter 14. Applications of interference cancellation and multi-user detection are addressed in Chapter 16. An overview of wired and wireless access technologies appears in Chapter 17, and specic examples of system design are given in Chapters 18 and 19.
Acknowledgements
We gratefully acknowledge all who have made the realization of this book possible. In particular, the editing of the various chapters would never have been completed without the contributions of numerous students in our courses on Algorithms for Telecommunications. Although space limitations preclude mentioning them all by name, we nevertheless express our sincere gratitude. We also thank Christian Bolis and Chiara Paci for their support in developing the software for the book, Charlotte Bolliger and Lilli M. Pavka for their assistance in administering the project, and Urs Bitterli and Darja Kropaci for their help with the graphics editing. For text processing of the Italian version, the contribution of Barbara Sicoli was indispensable; our thanks also go to Jane Frankeneld Zanin for her help in translating the text into English. We are pleased to thank the following colleagues for their invaluable assistance throughout the revision of the book: Antonio Assalini, Paola Bisaglia, Alberto Bononi, Giancarlo Calvagno, Giulio Colavolpe, Roberto Corvaja, Elena Costa, Andrea Galtarossa, Antonio Mian, Carlo Monti, Ezio Obetti, Riccardo Rahely, Roberto Rinaldo, Antonio Salloum, Fortunato Santucci, Andrea Scaggiante, Giovanna Sostrato, Stefano Tomasin, and Luciano Tomba. We gratefully acknowledge our colleague and mentor Jack Wolf for letting us include his lecture notes in the chapter on channel codes. A special acknowledgment goes to our colleagues Werner Bux and Evangelos Eleftheriou of the IBM Zurich Research Laboratory, and Silvano Pupolin of the University of Padua, for their continuing support. Nevio Benvenuto Giovanni Cherubini
To make the reading of the adopted symbols easier, a table containing the Greek alphabet is included. The Greek alphabet A B 0 , " 1 E Z , # H 2 I K 3 M alpha beta gamma delta epsilon zeta eta theta iota kappa lambda mu ! o , % , & ,' N 4 O 5 P P T Y 8 X 9 nu xi omicron pi rho sigma tau upsilon phi chi psi omega
Chapter 1
In the present chapter we recall fundamental concepts of signal theory and random processes. A majority of readers will simply nd this chapter a review of known principles, while others will nd it a useful incentive for further in-depth study, for which we recommend the items in the bibliography. In any event, we will begin with the denition of signal space and its discrete representation, then move to the study of discrete-time linear systems (discrete Fourier transforms, IIR and FIR impulse responses) and signals (complex representation of passband signals and the baseband equivalent). We will conclude with the study of random processes, with emphasis on the statistical estimation of rst- and second-order ergodic processes (periodogram, correlogram, ARMA, MA and especially AR models).
1.1
Signal space
Denition 1.1 A linear space is a set of elements called vectors, together with two operators dened over the elements of the set, the sum between vectors and the multiplication of a vector by a scalar. The Euclidean space is an example of linear space in which the sum of two vectors coincides with the vector obtained by adding the individual components, and the product of a vector by a scalar coincides with the vector obtained by multiplying each component for that scalar. In our case of particular interest is the set of complex vectors, i.e., those with complex-valued components, in an Euclidean space.
(1.5)
(1.6)
(1.7) (1.8)
A geometrical interpretation of the two elementary operations in a two-dimensional Euclidean space is given in Figure 1.1. As previously mentioned, the Euclidean space is an example of a linear space. Two other examples of linear spaces are: the discrete-time signal space (an Euclidean space with innite dimensions), whose elements are the signals fx.kTc /g k integer (1.9)
where Tc is the sampling period or interval,1 and the continuous-time signal space, whose elements are the signals x.t/ where < denotes the set of real numbers. t 2< (1.10)
Figure 1.1. Geometrical interpretation in the two-dimensional space of the sum of two vectors and the multiplication of a vector by a scalar.
Later a discrete-time signal will be indicated simply as fx.k/g, omitting the indication of the sampling period. In general, we will indicate by fxk g a sequence of real or complex numbers not necessarily generated at instants kTc .
Inner product
In an I -dimensional Euclidean space,2 given the two vectors x D [x1 ; : : : ; x I ]T and y D [y1 ; : : : ; y I ]T , we indicate with hx; yi the inner product: hx; yi D
I X i D1
xi yi
(1.11)
If hx; yi is real, there is an important geometrical interpretation of the inner product in the Euclidean space, represented in Figure 1.2, that is obtained from the relation: hx; yi D jjxjj jjyjj cos where jjxjj denotes the norm or length of the vector x. Note that hx; xi D
I X i D1
(1.12)
jxi j2 D jjxjj2
(1.13)
Observation 1.1 From (1.12), hx; yi D jjxjj cos jjyjj is the length of the projection of x onto y. Denition 1.2 Two vectors x and y are orthogonal (x ? y) if hx; yi D 0, that is if the angle they form is 90 .
(I=2)
(1.14)
y ||y|| x ||x||
Figure 1.2. Geometrical representation of the inner product between two vectors. jjxjj is the norm of x, that is the vector length.
Henceforth: T stands for transpose, for complex conjugate and H for transpose complex conjugate or Hermitian.
We can extend these concepts to a signal space, dening the inner product as
C1 X kD 1
x.k/ y .k/
(1.15)
x.t/ y .t/ dt
(1.16)
for continuous-time signals. In both cases it is assumed that the energy of signals is nite. Hence, for continuous-time signals it must be: Z
C1 1
jx.t/j2 dt < 1
Z and
C1 1
jy.t/j2 dt < 1
(1.17)
Recall that the inner product enjoys the following properties: 1. hx C y; zi D hx; zi C hy; zi. 2. hx; yi D hx; yi. 3. hx; yi D hy; xi . 4. hx; xi > 0 8x 6D 0.
5. (Schwarz inequality) jhx; yij jjxjjjjyjj. Equality holds if and only if x D Ky, with K a complex scalar.
1.2
Let us consider the problem of associating a sequence (possibly nite) of numbers with a continuous-time signal.3 A basis of orthonormal signals (orthogonal signals with unit norm) f i .t/g, t 2 <, i 2 I, where I is a nite or numerable set, is dened by Z C1 1 if i D j (1.18) h i; ji D i .t/ j .t/ dt D i j D 0 if i 6D j 1 In this text, n is the Kronecker delta, whereas .t/ denotes the Dirac delta. Given a nite-energy signal x.t/, t 2 <, Z E x D hx; xi D
1 C1
jx.t/j2 dt < 1
(1.19)
we want to express x.t/, t 2 <, as a linear combination of the functions f i .t/g, i 2 I. Consider the signal X x.t/ D O xi i .t/ (1.20)
i 2I
the most common method to determine the coefcients fxi g in (1.20) is by minimizing the energy of e, dened as 2 Z C1 X E e D he; ei D xi i .t/ dt (1.22) x.t/ 1
i 2I
i 2I
(1.23)
hx; xi O X
i 2I
.xi xi
ci j2
X
i 2I
jci j2
(1.25)
as jxi j2 xi ci xi ci C jci j2 D jxi ci j2 D .xi ci /.xi ci / . The minimum of (1.25) is obtained if the second term is zero. Hence the coefcients to be determined are given by Z C1 x.t/ i .t/ dt i 2I (1.26) xi D ci D hx; i i D
1
and Ee D E x where Ex D O X
i 2I
Ex O jxi j2
(1.27)
(1.28)
If E e D 0, then f i .t/g, i 2 I, is a complete basis for x.t/, t 2 <, and x.t/ can be expressed as X xi i .t/ (1.29) x.t/ D
i 2I
where equality must be intended in terms of quadratic norm. Moreover, from (1.29) X jxi j2 (1.30) Ex D
i 2I
that is the energy of the signal coincides with the sum of the squares of the coefcient amplitudes.
D hx
x; O
ii
D xi
xi D 0
8i 2 I
(1.31)
that is e ? i , 8i 2 I. As an example, for a generic signal x and for a basis formed by two signals one gets the geometrical representation in Figure 1.3.
and
2,
Signal representation
Denition 1.3 The signal x.t/, t 2 <, given by (1.20) with fxi D ci g, i 2 I, is called the projection of O x.t/, t 2 <, onto the space spanned by the signals f i .t/g, i 2 I, that is, the space whose signals are expressed as linear combinations of f i .t/g, i 2 I. If E e D 0, then x.t/, t 2 <, belongs to the space spanned by f i .t/g, i 2 I. Therefore, given a sequence of orthonormal signals, which form a complete basis for x.t/, the signal x.t/ can be represented by a sequence of numbers fxi g, i 2 I, given by xi D hx;
ii
(1.32)
2 x e ^ x
1
Figure 1.3. Geometrical representation of the projection of x onto the space spanned by 1 and 2 .
In short, we have the following correspondence between a signal and its vector representation: x.t/ t 2< ! x D [: : : ; x
1 ; x 0 ; x 1 ; : : :] T
(1.33)
It is useful to analyze the inner product between signals in terms of the corresponding vector representations. Let X X x.t/ D xi i .t/ and y.t/ D yi i .t/ (1.34)
i 2I i 2I
then Z hx; yi D
1 C1
(1.35)
In particular, Z Ex D
1 C1
jx.t/j2 dt D
X
i 2I
(1.36)
D
1
jx.t/ !1
y.t/j dt
1
2
(1.37)
X
i 2I
jxi
yi j2
D d.x; y/
(1.38)
In other words, the Euclidean distance between two signals coincides with the Euclidean distance between the corresponding vectors. Moreover, the following relation holds:4 d 2 .x; y/ D E x C E y 2Re hx; yi (1.39) O In particular we have E e D d 2 .x; x/. Let D hx; yi jjxjj jjyjj (1.40)
be the correlation coefcient between x and y. If is real, observing (1.12) D cos , with the angle formed by x and y. Then (1.39) becomes: p d 2 .x; y/ D E x C E y 2 E x E y (1.41) We note that if is real, the signals are not necessarily real, it is sufcient that their inner product is real.
The symbols Re [c] and Im [c] denote, respectively, the real and the imaginary part of c.
Example 1.2.1 Resorting to the sampling theorem, it can be shown that for a real valued signal x.t/, t 2 <, with nite bandwidth B (see (1.140)), the sequence of functions 1 sin .t i Tc / 1 Tc (1.42) Tc D i .t/ D 1 2B .t i Tc / Tc forms a complete orthogonal basis for x. The coefcients are given by the samples of x.t/ at the time instants t D i Tc , xi D x.i Tc / (1.43)
a procedure to derive an orthonormal basis for this set is now outlined. We indicate by f i0 .t/g a set of orthogonal functions and by 8 9 < 0 .t/ = .t/ D qi (1.45) : i E i0 ; the orthonormal functions obtained from f i0 .t/g. 1. Let
0 1 .t/
D s1 .t/
(1.46)
Then it follows
1 .t/ 0 .t/ D q1 0 E 1
(1.47)
2. Let
0 2 .t/
D s2 .t/
hs2 ;
1 i 1 .t/
(1.48)
0 2
1;
in fact, hs2 ;
1i 1; 1i
D hs2
D hs2 ;
1i
hs2 ;
1i
D0
(1.49)
1.
1 i 1 .t/
0 .t/ D q2 0 E 2
(1.50)
3. Let
0 3 .t/
D s3 .t/
0 3
hs3 ;
2.
1 i 1 .t/
hs3 ;
2 i 2 .t/
(1.51)
0 3
and
0 i .t/
D si .t/
i 1 X jD1
hsi ;
ji
j .t/
(1.52)
and if
0 i .t/
Observation 1.2 The set of f i .t/g is not unique, in any case the reciprocal distances between signals remain unchanged. Observation 1.3 The number of dimensions I of f i .t/g can be lower than M if the signals fsm .t/g, m D 1; : : : ; M, are linearly dependent, that is if there exists a set of coefcients, not all equal to zero, such that
M X mD1
cm sm .t/ D 0
8t
(1.54)
In such a case, it happens that in (1.52) for some i the signal f i0 .t/g is identically zero. Obviously, a null signal cannot be an element of the basis. Let us look at a few examples of discrete representation of a set of signals.
10
Example 1.2.2 For the three signals 8 < A sin 2 t T s1 .t/ D : 0 8 < A sin 2 t T s2 .t/ D : 0 8 < A sin 2 t T s3 .t/ D : 0 T 2 elsewhere 0<t < 0<t <T elsewhere T <t <T 2 elsewhere
(1.55)
(1.56)
(1.57)
which are depicted in Figure 1.5, an orthonormal basis, represented in Figure 1.6, is the following: 8 T 2 < p sin 2 t 0<t < T 2 (1.58) T 1 .t/ D : 0 elsewhere 8 2 T < p sin 2 t <t <T T 2 (1.59) T 2 .t/ D : 0 elsewhere
s1(t)
s (t)
0
T
s (t)
11
_ _ 2/ \|T
_ _ 2/ \|T
1(t)
2(t)
_ _ 2/ \|T
0
_ _ 2/ \|T
T
(1.60) Ap T 2
2 .t/
(1.61) (1.62)
from which the correspondence between signals and their vector representation is (see Figure 1.7): T Ap s1 .t/ ! s1 D T;0 (1.63) 2 Ap Ap T s2 .t/ ! s2 D T; T (1.64) 2 2 Ap T s3 .t/ ! s3 D 0; T (1.65) 2
2 s3 A T 2 s2
s1 0 A T 2 1
12
We note that the three signals are represented as a linear combination of only two functions (I D 2). Denition 1.4 The vector representation of a set of M signals is often called a signal constellation. Example 1.2.3 (4-PSK) Given the set of four signals 8 < A cos 2 f 0 t C m sm .t/ D : 0
1 2
m D 1; 2; 3; 4
(1.66)
depicted in Figure 1.8, to determine the basis functions we write sm .t/ as 1 1 sm .t/ D A cos m cos.2 f 0 t/ A sin m sin.2 f 0 t/ 2 2 2 2 We choose the following basis: 8 r > < C 2 cos.2 f t/ 0 T 1 .t/ D > : 0
(1.67)
s (t)
s (t)
0
T/2
T/2
s (t)
s (t)
0
T/2
T/2
13
2 s2 A 2 s1 T
A 2
s3
s4
and
2 .t/
2 sin.2 f 0 t/ T
(1.68)
D1C
sin 4 f 0 T 4 f0 T
D1
sin 4 f 0 T 4 f0 T
(1.69)
sin2 2 f 0 T 2 f 0 T
(1.70)
then it results h 1 ; 2 i ' 0, and E i ' 1 for i D 1; 2. Under these conditions for f 0 , 1 .t/ and 2 .t/ form an othonormal basis for the four signals in Figure 1.8, whose constellation is given in Figure 1.9.
1.3
A time-invariant continuous-time continuous-amplitude linear system, also called analog lter, is represented in Figure 1.10, where x and y are the input and output signals, respectively, and h denotes the lter impulse response.
14
x(t)
y(t)
Figure 1.10. Analog lter as a time-invariant linear system with continuous domain.
The output at a certain instant t 2 < is given by the convolution integral Z 1 Z 1 y.t/ D h.t / x. / d D h. / x.t / d
1 1
(1.72)
In short we will use the notation y.t/ D x h.t/ where the symbol denotes the convolution operation (1.72). We also introduce the Fourier transform of the signal x.t/, t 2 <, Z C1 X . f / D F[x.t/] D x.t/ e j2 ft dt f 2<
1
(1.73)
(1.74)
(1.75)
where H is the lter frequency response. The magnitude of the frequency response, jH. f /j, is usually called the magnitude response or amplitude response. General properties of the Fourier transform are given in Table 1.1.5 Denition 1.5 We introduce two functions that will be extensively used: 8 F < 1 jfj< f 2 D rect : F 0 elsewhere
(1.77)
Two important functions that will be used very often are: ( step function: 1.t/ D ( sign function: sgn.t/ D 1 0 1 1 t >0 t <0 t >0 t <0
15
Table 1.1 Some general properties of the Fourier transform. Property Signal x.t/ linearity duality time inverse complex conjugate real part imaginary part scaling time shift frequency shift modulation a x.t/ C b y.t/ X .t/ x. t/ x .t/ Re[x.t/] D Im[x.t/] D x.t/ C x .t/ 2 x.t/ x .t/ 2j Fourier transform X. f / a X . f / C b Y. f / x. X. X . f/ f/ f/
x.at/, a 6D 0 x.t t0 /
1 [X . f / C X . f /] 2 1 [X . f / X . f /] 2j f 1 X jaj a e j2 f t0 X . f / X. f f0 /
1 j' f 0 / C e j' X . f C f 0 /] [e X . f 2 1 f 0 / e j' X . f C f 0 /] [e j' X . f 2j 1 j' f 0 / C e j' X . f f 0 /] [e X . f 2 j2 f X . f / 1 X .0/ . f / X. f / C 2 j2 f X . f / Y. f / X . f / Y . f / [X ./ Y./]. f / X . f / D X . f /, X Hermitian, Re[X . f /] even, Im[X . f /] odd, jX . f /j2 even X. f / D X . f/ f /, X real and even X. f /, X . f / D X . f / D X . X. f / D Z C1 X . f / D X imaginary and odd
differentiation integration convolution correlation product real signal imaginary signal real and even signal real and odd signal
d x.t/ dt Z t x. / d D 1 x.t/
1
[x. / y. /].t/ [x. / y . /].t/ x.t/ y.t/ x.t/ D x .t/ x.t/ D x .t/ x. t/
Ex D
jX . f /j2 d f D E X 1 ` x.kTc / D X Tc `D 1 Tc kD 1
1 C1 X
jx.t/j2 dt D
16
sin. t/ t
(1.78)
f F
(1.79)
Further examples of signals and relative Fourier transforms are given in Table 1.2. We reserve the notation H .s/ to indicate the Laplace transform of h.t/, t 2 <: H .s/ D Z
C1
h.t/e
1
st
dt
(1.80)
with s complex variable; H .s/ is also called the transfer function of the lter. A class of functions H .s/ often used in practice is characterized by the ratio of two polynomials in s, each with a nite number of coefcients. It is easy to observe that if the curve s D j2 f in the s-plane belongs to the convergence region of the integral in (1.80), then H. f / is related to H .s/ by H. f / D H .s/jsD j2 f
sinc(tF) 1
(1.81)
4/F
3/F
2/F
1/F
1/F
2/F
3/F
4/F
1/Frect(f/F)
1/F
F/2
F/2
17
Table 1.2 Examples of Fourier transform signal pairs. Signal x.t/ .t/ 1 e j2 f 0 t cos.2 f 0 t/ sin.2 f 0 t/ 1.t/ sgn.t/ t rect T t sinc T t jtj 1 rect T 2T e at 1.t/, a > 0 t e at 1.t/, a > 0 e ajtj , a > 0
2 e at , a > 0
1.4
A discretetime time-invariant linear system, with sampling period Tc , is shown in Figure 1.12, where x.k/ and y.k/ are respectively the input and output signals at the time instants kTc , k 2 Z, where Z denotes the set of integers. The impulse response of the system is denoted by fh.k/g, k 2 Z, or more simply by h.
x(k) Tc y(k) Tc
18
The relation between the input sequence fx.k/g and the output sequence fy.k/g is given by the convolution operation: y.k/ D [x.m/ h.m/].k/ D
C1 X nD 1
h.k
n/x.n/
(1.82)
In short, we will use the notation y.k/ D x h.k/. We list some denitions that are valid for time-invariant linear systems. We say the system is causal (anticausal ) if h.k/ D 0, k < 0 (if h.k/ D 0, k > 0). We dene as transfer function of the lter the z-transform6 of the impulse response h, given by H .z/ D
C1 X kD 1
h.k/z
(1.83)
h.k/e
j2 f kTc
D H .z/ zDe j2 f Tc
(1.84)
H. f /e j2 f kTc d f
(1.85)
We note the property that, for x.k/ D bk , where b is a complex constant, the output is given by y.k/ D H .b/ bk . In Table 1.3 some further properties of the z-transform are summarized. For discrete-time linear systems, in the frequency domain (1.82) becomes Y. f / D X . f /H. f / where all functions are periodic of period 1=Tc . Example 1.4.1 A fundamental example of ztransform is that of the sequence: ( ak k0 jaj < 1 h.k/ D 0 k<0 Applying the transform formula (1.83) we nd H .z/ D under the condition ja=zj < 1.
6
(1.86)
(1.87)
1 az
z z a
(1.88)
Sometimes the D transform is used instead of the z-transform, where D D z 1 , and H .z/ is replaced by P h.D/ D C1 1 h.k/D k . kD
19
Property
linearity
ax.k/ C by.k/
delay x.k m/ complex conjugate x .k/ inverse time x. k/ x . k/ scaling convolution correlation real sequence a
k x.k/
X .z/Y .z/ 1 [x.m/ y . m/].k/ X .z/Y z X .z/ D X .z / x.k/ D x .k/ [x.m/ y.m/].k/
Example 1.4.2 Let q.t/, t 2 <, be a continuous-time signal with Fourier transform Q. f /, f 2 <. We now consider the sequence obtained by sampling q.t/, that is h k D q.kTc / k2Z (1.89)
Using the Poisson formula of Table 1.1, one demonstrates that the Fourier transform of the sequence fh k g is related to Q. f / by 1 1 X 1 (1.90) Q f l H. f / D F[h k ] D H e j2 f Tc D Tc `D 1 Tc
1, the expression
gk e
j2 f kTc
WN D e
2 j N
(1.92)
20
1.
X 1 N 1 Gm W N km N mD0
k D 0; 1; : : : ; N
(1.93)
We note that, besides the factor 1=N , the expression of the inverse DFT (IDFT) coincides with that of the DFT, provided W N 1 is substituted with W N . We also observe that direct computation of (1.92) requires N .N 1/ complex additions and N 2 complex multiplications; however, the algorithm known as fast Fourier transform (FFT) allows computation of the DFT by N log2 N complex additions and N log2 N N 2 complex multiplications.7
3 7 7 7 7 7 7 7 7 5
(1.94)
p We note that F D FT , and .1= N /F is a unitary matrix.8 The following property holds: if C is a right circulant square matrix whose rows are obtained by successive shift to the right of the rst row, then FCF 1 is a diagonal matrix whose elements are given by the DFT of the rst row of C. Introducing the vector formed by the samples of the sequence fgk g, k D 0; 1; : : : ; N 1, gT D [g0 ; g1 ; : : : ; g N and the vector of transform coefcients G T D [G0 ; G1 ; : : : ; G N G D Fg
1] 1]
(1.96)
D DFT[g]
(1.97)
7 8
The computational complexity of the FFT is often expressed as N log2 N . A square matrix A is unitary if A H A D I where I is the identity matrix, i.e. a matrix for which all elements are zero except the elements on the main diagonal that are all equal to one.
21
k<0
k > Lx
(1.100)
k>N
(1.101)
We dene the periodic signals of period L, xrep L .k/ D and h rep L .k/ D
C1 X `D 1
x.k
`L/
(1.102)
h.k
`L/
(1.103)
Denition 1.6 The circular convolution between x and h is a periodic sequence of period L dened as y .circ/ .k/ D h
L
x.k/ D
L 1 X i D0
i/
(1.105)
Lx -1
0 1 N-1
k
1.
1, and fh(k)g, k D 0; 1; : : : ; N
22
.circ/ Then, if we indicate with fXm g, fHm g, and fYm g, m D 0; 1; : : : ; L .circ/ , respectively, we obtain DFT of sequences x, h, and y .circ/ Ym D Xm Hm
m D 0; 1; : : : ; L
In vector notation (1.97), (1.106) becomes9 i h .circ/ .circ/ .circ/ T Y .circ/ D Y0 ; Y1 ; : : : ; Y L 1 D diagfDFT[x]gH
(1.107)
where H is the column vector given by the L-point DFT of the sequence h. We are often interested in the linear convolution between x and h given by (1.82): y.k/ D x h.k/ D whose support is k D 0; 1; : : : ; L x C N to see that only if
N 1 X i D0
h.i/x.k
i/
(1.108)
k D 0; 1; : : : ; L
(1.110)
To compute the convolution between the two nite-length sequences x and h, (1.109) and (1.110) require that both sequences be completed with zeros (zero padding) to get a length of L D L x C N 1 samples. Then, taking the L-point DFT of the two sequences, performing the product (1.106), and taking the inverse transform of the result, one obtains the desired linear convolution. We give below two relations between the circular convolution y .circ/ and the linear convolution y; in both cases we use L D Lx with L > N . Relation 1. We verify that the two convolutions y .circ/ and y coincide only for the instants k D N 1; N ; : : : ; L 1, and we write y .circ/ .k/ D y.k/ only for k D N 1; N ; : : : ; L 1 (1.112) (1.111)
Indeed, with reference to Figure 1.14, the result of circular convolution coincides with fy.k/g, output of the linear convolution, only for a delay k such that it is avoided the product between non-zero samples of the two periodic sequences h r ep L and xr ep L , indicated by and , respectively. This is achieved only for k N 1 and k L 1.
The notation diagfvg denotes a diagonal matrix whose elements on the diagonal are equal to the components of the vector v.
23
hrep L (i)
xrep L (k-i)
N-1 L-1
k-(L-1)
Figure 1.14. Illustration of the circular convolution operation between fx(k)g, k D 0; 1; : : : ; L 1, and fh(k)g, k D 0; 1; : : : ; N 1.
Relation 2. An alternative to (1.112) is to consider, instead of the nite sequence x, an extended sequence x . px/ that is obtained by partially repeating x with a cyclic prex of N px samples: ( x.k/ k D 0; 1; : : : ; L x 1 . px/ x .k/ D (1.113) k D N px ; : : : ; 2; 1 x.L x C k/ Let y . px/ be the linear convolution between x . px/ and h, with support f N px ; : : : ; L x C N 2g. If N px N 1, it is easy to prove the following relation y . px/ .k/ D y .circ/ .k/ Let us dene z.k/ D ( y . px/ .k/ 0 k D 0; 1; : : : ; L x k D 0; 1; : : : ; L x elsewhere 1 (1.114)
(1.115)
then from (1.114) and (1.106) the following relation between the corresponding L x point DFTs is obtained: Zm D Xm Hm m D 0; 1; : : : ; L x 1 (1.116)
10
24
h x
0T 0T
(1.117) (1.118)
in which we have assumed x.k/ D 0, k D 0; 1; : : : ; N 2. Transform H0 D DFT[h0 ] X 0 D diagfDFT[x 0 ]g 3. Matrix product Y 0 D X 0 H0 4. Inverse transform h i z }| { D [ ]; : : : ; ] ; y.N
N 1 terms
vector matrix
(1.119) (1.120)
vector
(1.121)
0T
D DFT
0T
1/]
(1.122)
where the symbol ] denotes a component that is neglected. The second loading contains x 0T D [x..L 1/ .N 2//; : : : ; x.2.L 1/ .N 2//] (1.123)
and the desired output samples will be y.k/ The third loading contains x 0T D [x.2.L 1/ 2.N 2//; : : : ; x.3.L 1/ 2.N 2//] (1.125) k D L ; : : : ; 2.L 1/ .N 2/ (1.124)
and will yield the desired output samples y.k/ k D 2.L 1/ .N 2/ C 1; : : : ; 3.L 1/ 2.N 2/ (1.126)
25
an y.k
n/ D
q X nD0
bn x.k
n/
(1.127)
where we can set a0 D 1 without loss of generality. If the system is causal, (1.127) becomes y.k/ D
p X nD1
an y.k
n/ C
q X nD0
bn x.k
n/
k0
(1.128)
and the transfer function for such systems assumes the form
q X
bn z
b0 D
n
q Y
.1
zn z pn z
1
/ (1.129)
nD0
1C
p X nD1
an z
nD1 p Y nD1
.1
where fzn g and fpn g are, respectively, the zeros and poles of H .z/. Equation (1.129) generally denes an innite impulse response (IIR) lter. In the case in which an D 0, n D 1; 2; : : : ; p, (1.129) reduces to H .z/ D
q X nD0
bn z
(1.130)
and we obtain a nite impulse response (FIR) lter with h.n/ D bn , n D 0; 1; : : : ; q. To get the impulse response coefcients, assuming known the z-transform H .z/, we can expand H .z/ in partial fractions and apply the linear property of the z-transform (see Table 1.3, page 19). If q < p and assuming that all poles are distinct, we obtain
p X nD1
H .z/ D
rn pn z
H) h.k/ D
k rn pn
k0 k<0
(1.131)
where rn D H .z/[1 We give now two denitions. Denition 1.7 A causal system is stable (bounded input-bounded output stability) if jpn j < 1, 8n. pn z
1
]jzDpn
(1.132)
26
Denition 1.8 The system is minimum phase (maximum phase) if jpn j < 1 and jzn j 1 (jpn j > 1 and jzn j > 1), 8n. Among all systems having the same magnitude response jH.e j2 f Tc /j, the minimum (maximum) phase system presents a phase response, argH.e j2 f Tc /, which is below (above) the phase response of all other systems. Example 1.4.3 It is interesting to determine the phase of a system for a given impulse response. Let us consider the system with transfer function H1 .z/ and impulse response h 1 .k/ shown in Figure 1.15a. After determining the zeros of the transfer function, we factorize H1 .z/ as: H1 .z/ D b 0
4 Y nD1
.1
zn z
(1.133)
As shown in Figure 1.15a, H1 .z/ is minimum phase. We now observe that the magnitude of the frequency response does not change if 1=z n is replaced with z n in (1.133). If we move all the zeros outside the unit circle, we get a maximum-phase system H2 .z/ whose impulse response is shown in Figure 1.15b. A general case, that is a transfer function with some zeros inside and others outside the unit circle, is given in Figure 1.15c. The coefcients of the impulse responses h 1 , h 2 , and h 3 are given in Table 1.4. The coefcients are normalized so that the three impulse responses have equal energy. We dene the partial energy of a causal impulse response as E.k/ D
k X i D0
jh.i/j2
(1.134)
Comparing the partial-energy sequences for the three impulse responses of Figure 1.15, one nds that the minimum (maximum) phase system yields the largest (smallest) fE.k/g. In other words, the magnitude of the frequency responses being equal, a minimum (maximum) phase system concentrates all its energy on the rst (last) samples of the impulse response. Extending our previous considerations also to IIR lters, if h 1 is a causal minimumphase lter, i.e. H1 .z/ D Hmin .z/ is a ratio of polynomials in z 1 with poles and zeros
Table 1.4 Impulse responses of systems having the same magnitude of the frequency response.
h.2/ 0 0 0:15e
j1:66
0:3e j0:63
j1:57
0:9e j1:57
j0:63
27
Figure 1.15. Impulse response magnitudes and zero locations for three systems having the same frequency response magnitude.
inside the unit circle, then Hmax .z/ D K Hmin z1 , where K is a constant, is an anticausal maximum-phase lter, i.e. Hmax .z/ is a ratio of polynomials in z with poles and zeros outside the unit circle. In the case of a minimum-phase FIR lter with impulse response h min .n/, n D 0; 1; : : : ; q, H2 .z/ D z q Hmin z1 is a causal maximum-phase lter. Moreover, the relation fh 2 .n/g D
B fh .q n/g, n D 0; 1; : : : ; q, is satised. In this text we use the notation fh 2 .n/g D fh 1 .n/g, 1 where B is the backward operator that orders the elements of a sequence from the last to the rst.
28
In Appendix 1.A multirate transformations for systems are described, in which the time domain of the input is different from that of the output. In particular, decimator and interpolator lters are introduced, together with their efcient implementations.
1.5
Signal bandwidth
Denition 1.9 The support of a signal x. /, 2 <, is the set of values 2 < for which jx. /j 6D 0. Let us consider a lter with impulse response h and frequency response H. If h assumes real values, then H is Hermitian, H. f / D H . f /, and jH. f /j is an even function. Depending on the support of jH. f /j, the classication of Figure 1.16 is usually done. If h assumes complex values, the terminology is less standard. We adopt the classication of Figure 1.17, in which the lter is a lowpass lter (LPF) if the support jH. f /j includes the origin, otherwise it is a passband lter (PBF).
Figure 1.16. Classication of real valued analog lters on the basis of the support of jH.f/j.
29
Figure 1.17. Classication of complex valued analog lters on the basis of support of jH.f/j.
Analogously, for a signal x, we will use the same denomination and we will say that x is a baseband (BB) or passband (PB) signal depending on whether the support of jX . f /j, f 2 <, includes or not the origin. Denition 1.10 In general, for a real-valued signal x, the set of positive frequencies such that jX . f /j 6D 0 is called passband or simply band B: B D f f 0 : jX . f /j 6D 0g (1.135)
As jX . f /j is an even function, we have jX . f /j 6D 0, f 2 B. We note that B is equivalent to the support of X limited to positive frequencies. The bandwidth11 of x is given by the measure of B: Z df (1.140) BD
B
11 The signal bandwidth may be given different denitions. Let us consider an LPF having frequency response
H. f /. The lter gain H0 is usually dened as H0 D jH.0/j; other denitions are as average gain of the lter in the passband B, or as max f jH. f /j. We give the following four denitions for the bandwidth B of h. a) First zero: B D minf f > 0 : H. f / D 0g (1.136)
30
For example, with regard to the signals of Figure 1.16, we have that for an LPF B D f 2 , whereas for a PBF B D f 2 f 1 . In the case of a complex-valued signal x, B is equivalent to the support of X , and B is thus given by the measure of the entire support. For discrete-time lters, for which H is periodic of period 1=Tc , the same denitions hold, with the caution of considering the support of jH. f /j within a period, lets say between 1=.2Tc / and 1=.2Tc /. In the case of discrete-time highpass lters (HPF), the passband will extend from a certain frequency f 1 to 1=.2Tc /. As discrete-time signals are often obtained by sampling continuous-time signals, we will state the following fundamental theorem.
univocally represent the signal q.t/, t 2 <, under the condition that the sampling frequency 1=Tc satises the relation 1 B0 Tc (1.142)
For the proof, which is based on the relation (1.90) between a signal and its samples, we refer the reader to [1]. B0 is often referred to as the minimum sampling frequency. If 1=Tc < B0 the signal cannot be perfectly reconstructed from its samples, originating the so-called aliasing phenomenon in the frequency-domain signal representation.
b) Based on amplitude, bandwidth at A dB: A jH. f /j B D max f > 0 : D 10 20 H0 Typically A D 3; 40, or 60. c) Based on energy, bandwidth at p%: Z B Z 01
0
(1.137)
p 100
(1.138)
Figure 1.18 illustrates the various denitions for a particular jH. f /j.
31
10
B3dB Breq
20
30
|H(f)| (dB)
40
50
60
B40dB
70
BE (p=90)
80
B50dB
90
BE (p=99)
100
0.2
0.4
0.6
0.8
1 f (Hz)
1.2
1.4
1.6
1.8
Figure 1.18. The real signal bandwidth following the denitions of: 1) Bandwidth at rst zero: Bz D 0:652 Hz. 2) Amplitude-based bandwidth: B3 dB D 0:5 Hz, B40 dB D 0:87 Hz, B50 dB D 1:62 Hz. 3) Energy-based bandwidth: BE.pD90/ D 1:362 Hz, BE.pD99/ D 1:723 Hz. 4) Equivalent noise bandwidth: Breq D 0:5 Hz.
In turn, the signal q.t/, t 2 <, can be reconstructed from its samples fh k g according to the scheme of Figure 1.19, where it is employed as an interpolation lter having an ideal frequency response given by ( GI . f / D 1 0 f 2B elsewhere
(1.143)
We note that for real-valued baseband signals B0 D 2B. For passband signals, care must be taken in the choice of B0 2B to avoid aliasing between the positive and negative frequency components of Q. f /.
32
(1.149)
We emphasize that it is sufcient that the Heaviside conditions are veried within the support of X ; as jX . f /j D 0 outside the support, the lter frequency response may be arbitrary. We show in Figure 1.20 the frequency response of a PBF, with bandwidth B D f 2 f 1 , that satises the conditions stated by Heaviside.
Figure 1.20. Characteristics of a lter satisfying the conditions for the absence of signal distortion in the frequency interval (f1 ; f2 ).
12 For a complex number c, arg c denotes the phase of c (see note 3, page 441).
33
1.6
Passband signals
Complex representation
For a passband signal x it is convenient to introduce an equivalent representation in terms of a baseband signal x .bb/ . Let x be a PB real-valued signal with Fourier transform as illustrated in Figure 1.21. The following two procedures can be adopted to obtain x .bb/ . PB lter. Referring to Figure 1.21 and to the transformations illustrated in Figure 1.22, given x we extract its positive frequency components using an analytic lter or phase splitter, h .a/ , having the following ideal frequency response ( 2 f >0 .a/ H . f / D 2 1. f / D (1.150) 0 f <0 In practice, it is sufcient that h .a/ be a complex PB lter, with passband, in which H.a/ . f / ' 2, that extends from f 1 to f 2 , equal to that of X . f /, and stopband, in which jH.a/ . f /j ' 0, that extends from f 2 to f 1 . The signal x .a/ is called the analytic signal or pre-envelope of x. It is now convenient to introduce a suitable frequency f 0 , called reference carrier frequency, which usually belongs to the passband . f 1 ; f 2 / of x. The lter output, x .a/ , is
Figure 1.21. Illustration of transformations to obtain the baseband equivalent signal x.bb/ around the carrier frequency f0 using a phase splitter.
34
-j2 f 0 t
Figure 1.22. Transformations to obtain the baseband equivalent signal x.bb/ around the carrier frequency f0 using a phase splitter.
frequency shifted by f 0 to obtain a BB signal, x .bb/ . The signal x .bb/ is the baseband equivalent of x, also called complex envelope of x around the carrier frequency f 0 . Analytically, we have x .a/ .t/ D x h .a/ .t/ x .bb/ .t/ D x .a/ .t/ e and in the frequency domain X
.bb/ j2 f 0 t
F F
(1.151) (1.152)
.f/ D
2X . f C f 0 / 0
(1.153)
In other words, x .bb/ is given by the components of x at positive frequencies, scaled by 2 and frequency shifted by f 0 . BB lter. One gets the same result using a frequency shift of x followed by a lowpass lter (see Figures 1.23 and 1.24). It is immediate to determine the relation between the frequency responses of the lters of Figure 1.21 and Figure 1.23: H. f / D H.a/ . f C f 0 / (1.154) From (1.154) one can derive the relation between the impulse response of the analytic lter and the impulse response of the lowpass lter: h .a/ .t/ D h.t/ e j2 f 0 t (1.155)
f / D X . f /1. f / C X .
f /1.
f/
(1.156)
(1.157) (1.158)
35
Figure 1.23. Illustration of transformations to obtain the baseband equivalent signal x.bb/ around the carrier frequency f0 using a lowpass lter.
LPF x(t) h e
-j2 f 0 t
x(bb)(t)
Figure 1.24. Transformations to obtain the baseband equivalent signal x.bb/ around the carrier frequency f0 using a lowpass lter.
x (bb) (t)
x (a) (t)
Re[ . ]
x(t)
e j2 f0 t
Figure 1.25. Relation between a signal, its complex envelope and the analytic signal.
36
where
.bb/ x I .t/ D Re[x .bb/ .t/]
(1.160)
and
.bb/ x Q .t/ D Im[x .bb/ .t/]
(1.161)
are real-valued baseband signals, called in-phase and quadrature components of x, respectively. Substituting (1.159) in (1.158) we obtain
.bb/ x.t/ D x I .t/ cos 2 f 0 t .bb/ x Q .t/ sin 2 f 0 t
(1.162)
as illustrated in Figure 1.26. Conversely, given x, one can use the scheme of Figure 1.24 and the relations (1.160) and (1.161) to get the baseband components. If the frequency response H. f / has Hermitiansymmetric characteristics with respect to the origin, h is real and the scheme of Figure 1.27a holds. The scheme of Figure 1.27b employs instead an ideal Hilbert lter with frequency response given by H.h/ . f / D j sgn. f / D e
j 2 sgn. f /
(1.163)
Magnitude and phase of H.h/ . f / are shown in Figure 1.28. We note that h .h/ phase-shifts by =2 the positive-frequency components of the input and by =2 the negative-frequency components. In practice these lter specications are imposed only on the passband of the input signal.13 To simplify the notation, in block diagrams a Hilbert lter is indicated as =2.
cos(2 f 0 t) x (bb) (t)
I
-sin(2 f 0 t)
Figure 1.26. Relation between a signal and its baseband components.
13 We note that the ideal Hilbert lter in Figure 1.28 has an impulse response given by (see Table 1.2 on page 17):
h .h/ .t/ D
1 t
(1.164)
37
Comparing the frequency responses of the analytic lter (1.150) and of the Hilbert lter (1.163), we obtain the relation H.a/ . f / D 1 C jH.h/ . f / (1.167)
Consequently, if x is the input signal, the output of the Hilbert lter (also denoted as Hilbert transform of x) is Z C1 x. / 1 x .h/ .t/ D d (1.165) 1 t
38
| H (h)(f)| 1
Figure 1.28. Magnitude and phase responses of the ideal Hilbert lter.
Then, letting x .h/ .t/ D x h .h/ .t/ the analytic signal can be expressed as x .a/ .t/ D x.t/ C j x .h/ .t/ Consequently, from (1.152), (1.160) and (1.161):
.bb/ x I .t/ D x.t/ cos 2 f 0 t C x .h/ .t/ sin 2 f 0 t
(1.168)
(1.169)
(1.170) (1.171)
.bb/
x.t/ sin 2 f 0 t
as illustrated in Figure 1.27b.14 We note that in practical systems, transformations to obtain, e.g., the analytic signal, the complex envelope, or the Hilbert transform of a given signal, are implemented by
Moreover, noting that from (1.163) . j sgn f /. j sgn f / D 1, taking the Hilbert transform of a signal we get the initial signal with the sign changed. Then it results: Z C1 .h/ x . / 1 x.t/ D d (1.166) t 1
14 We recall that the design of a lter, and in particular of a Hilbert lter, requires the introduction of a suitable delay. In other words, we are only able to produce an output with a delay t D , x .h/ .t t D /. Consequently, in
the block diagram of Figure 1.27, also x.t/ and the various sinusoidal waveforms must be delayed.
39
ltering operations. However, it is usually more convenient to perform signal analysis in the frequency domain by the Fourier transform. In the following two examples we use frequency-domain techniques to obtain the complex envelope of a PB signal. Example 1.6.1 Let x.t/ be a sinusoidal signal, x.t/ D A cos.2 f 0 t C '0 / Then X. f / D A j'0 e . f 2 f0/ C A e 2
j'0
(1.172)
. f C f 0 /
(1.173)
The analytic signal is given by: X .a/ . f / D Ae j'0 . f and X .bb/ . f / D Ae j'0 . f /
F
1
f0 /
(1.174)
(1.175)
We note that we have chosen as reference carrier frequency of the complex envelope the same carrier frequency as in (1.172). Example 1.6.2 Let x.t/ D A sinc.Bt/ cos.2 f 0 t/ with the Fourier transform given by f f0 f C f0 A rect C rect X. f / D 2B B B as illustrated in Figure 1.29. Then, using f 0 as reference carrier frequency, f A .bb/ . f / D rect X B B and x .bb/ .t/ D A sinc.Bt/ (1.179) (1.176)
(1.177)
(1.178)
Another analytical technique to get the expression of the signal after the various transformations is obtained by applying the following theorem.
40
A 2B f 0 B B f0 f 0 + 2 2 A B
X (f)
0 X (bb)(f)
f0
f0
f0 +
B 2
B 2
f R
Theorem 1.1 Let the product of two real signals be x.t/ D a.t/ c.t/ (1.180)
where a is a BB signal with Ba D [0; B/ and c is a PB signal with Bc D [ f 0 ; C1/. If f 0 > B, then the analytic signal of x is related to that of c by: x .a/ .t/ D a.t/ c.a/ .t/ Proof. We consider the general relation (1.157), valid for every real signal c.t/ D Substituting (1.182) in (1.180) yields x.t/ D a.t/ 1 c.a/ .t/ C a.t/ 1 c.a/ .t/ 2 2 (1.183)
1 .a/ 1 .a/ .t/ 2 c .t/ C 2 c
In the frequency domain the support of the rst term in (1.183) is given by the interval [ f 0 B; C1/, while that of the second is equal to . 1; f 0 C B]. Under the hypothesis that f 0 B, the two terms in (1.183) have disjoint support in the frequency domain and (1.181) is immediately obtained. Corollary 1.1 From (1.181) we obtain x .h/ .t/ D a.t/c.h/ .t/ and x .bb/ .t/ D a.t/c.bb/ .t/ (1.185) (1.184)
B 2
B f R 2
(1.181)
(1.182)
41
In fact, from (1.169) we get x .h/ .t/ D Im[x .a/ .t/] (1.186)
which substituted in (1.181) yields (1.184). Finally, (1.185) is obtained by substituting (1.152), x .bb/ .t/ D x .a/ .t/e in (1.181). An interesting application of (1.186) is in the design of a Hilbert lter h .h/ starting from a lowpass lter h. In fact, from (1.155) and (1.186) we get h .h/ .t/ D h.t/ sin.2 f 0 t/ Example 1.6.3 Let a modulated double sideband (DSB) signal be expressed as x.t/ D a.t/ cos.2 f 0 t C '0 / (1.189) where a is a BB signal with bandwidth B. Then, if f 0 > B, from the above theorem we have the following relations: x .a/ .t/ D a.t/e j .2 f 0 tC'0 / x .h/ .t/ D a.t/ sin.2 f 0 t C '0 / x
.bb/ j2 f 0 t
(1.187)
(1.188)
.t/ D a.t/e
j'0
We list in Table 1.5 some properties of the Hilbert transformation (1.168) that are easily obtained by using the Fourier transform and the properties of Table 1.1.
Table 1.5 Some properties of the Hilbert transform.
Property
(Real) Hilbert transform x .h/ .t/ x.t/ x .h/ . t/ x .h/ .t/ D x .h/ .t/ D a.t/ c.h/ .t/ x .h/ . t/, odd x .h/ . t/, even
duality inverse time even signal odd signal product (see Theorem 1.1) cosinusoidal signal energy orthogonality
a.t/ c.t/ sin.2 f 0 t C '0 / cos.2 f 0 t C '0 / Z C1 Z C1 Ex D jx.t/j2 dt D jx .h/ .t/j2 dt D E x .h/
1 1
hx; x .h/ i D
C1 1
42
43
" D Re D Re hx h. / x
.bb/
. /
j'1
2 .t/
C h. / x
.bb/
eC j .2 2 f 0 C'1 / . / 2
# .t/ (1.193)
.bb/ e
j'1
where the last equality follows because the term with frequency components around 2 f 0 is ltered by the LPF. .bb/ We note, moreover, that the lter h .bb/ in Figure 1.30c has in-phase component h I .bb/ and quadrature component h Q that are related to H.a/ by (see (1.160) and (1.161)) H.bb/ . f / D 1 [H.bb/ . f / C H.bb/ . I 2 f /] f C f 0 /] (1.194)
f /] (1.195) . f C f 0 /]
1 [H.a/ . f C f 0 / D 2j
.a/
and
H.bb/ . f / D 0 Q
In other words h .bb/ .t/ D h .bb/ .t/ is real and the realization of the lter 1 h .bb/ is simplied. I 2 In practice this condition is veried by ensuring that the lter h .a/ has symmetrical frequency specications around f 0 .
44
In terms of the complex envelope signal x .bb/ , from (1.152) the equivalent relations follow: Mx .t/ D jx .bb/ .t/j 'x .t/ D arg x .bb/ .t/ C 2 f 0 t f x .t/ D 1 d [arg x .bb/ .t/] C f 0 2 dt (1.199) (1.200) (1.201)
Then, from the polar representation x .a/ .t/ D Mx .t/ e j'x .t/ and from (1.157), a PB signal x can be written as x.t/ D Re[x .a/ .t/] D Mx .t/ cos.'x .t// (1.202)
Two simplied methods to get the envelope Mx .t/ from the PB signal x.t/ are given in Figure 6.58 on page 514. For example if x.t/ D A cos.2 f 0 t C '0 / it follows that Mx .t/ D A 'x .t/ D 2 f 0 t C '0 f x .t/ D f 0 With reference to the above relations, three other denitions follow. 1. Envelope deviation 1Mx .t/ D jx .a/ .t/j 2. Phase deviation 1'x .t/ D 'x .t/ 3. Frequency deviation 1 f x .t/ D f x .t/ Then (1.202) becomes x.t/ D [A C 1Mx .t/] cos.2 f 0 t C '0 C 1'x .t// (1.209) f0 D 1 d 1'x .t/ 2 dt (1.208) .2 f 0 t C '0 / D arg x .bb/ .t/ '0 (1.207) A D jx .bb/ .t/j A (1.206) (1.203) (1.204) (1.205)
1.7
We recall the functions related to the statistical description of random processes, especially those functions concerning second-order analysis.
45
1.7.1
Correlation
Let x.t/ and y.t/, t 2 <, be two continuous-time random processes. We indicate the delay or lag with and the expectation operator with E. 1. Mean value mx .t/ D E[x.t/] 2. Statistical power Mx .t/ D E[jx.t/j2 ] 3. Autocorrelation rx .t; t 4. Cross-correlation rx y .t; t 5. Autocovariance cx .t; t / D E[.x.t/ D rx .t; t 6. Cross-covariance cx y .t; t / D E[.x.t/ D rx y .t; t Observation 1.4 x and y are orthogonal if rx y .t; t x and y are uncorrelated if cx y .t; t mx .t//.y.t / / m y .t / // ] (1.215) mx .t//.x.t / / mx .t / // ] (1.214) / D E[x.t/y .t /] (1.213) / D E[x.t/x .t /] (1.212) (1.211) (1.210)
mx .t/m .t x
mx .t/m .t y
if at least one of the two random processes has zero mean, orthogonality is equivalent to uncorrelation. x is wide-sense stationary (WSS) if 1. mx .t/ D mx , 8t, 2. rx .t; t / D rx . /, 8t.
15 We observe that the notion of orthogonality between two random processes is quite different from that of
orthogonality between two deterministic signals. In fact, while in the deterministic case, based on Denition 1.2, it is sufcient that the inner product of the signals is zero, in the random case the cross-correlation must be zero for all the delays and not only for zero delay. In particular, we note that the two random variables v1 and v2 are orthogonal if condition E[v1 v2 ] D 0 is satised.
46
2 rx .0/ D E[jx.t/j2 ] D Mx is the statistical power, whereas cx .0/ D x D Mx is the variance of x.t/.
jmx j2
x.t/ and y.t/ are jointly wide-sense stationary if 1. mx .t/ D mx , m y .t/ D m y , 8t, 2. rx y .t; t / D rx y . /, 8t.
1.7.2
Given the WSS random process x.t/, t 2 <, its power spectral density (PSD) is dened as the Fourier transform of the autocorrelation function Px . f / D F[rx . /] D Z
C1 1
rx . /e
j2 f
(1.216)
Px . f /e j2 f d f
(1.217)
Px . f / d f
(1.218)
hence the name PSD for the function Px . f /: it represents the distribution of the statistical power in the frequency domain. The pair of equations (1.216) and (1.217) are obtained from the WienerKhintchine theorem [2]. Denition 1.11 The passband B of a random process x is dened with reference to its PSD function.
47
(1.219)
X
i 2I
Mi . f
fi /
(1.220)
where I identies a discrete set of frequencies f f i g, i 2 I. The inverse Fourier transform of (1.219) yields the relation rx . / D r.c/ . / C r.d/ . / x x with r.d/ . / D x X
i 2I
(1.221)
Mi e j2 fi
(1.222)
The most interesting consideration is that the following random process decomposition corresponds to the decomposition (1.219) of the PSD: x.t/ D x .c/ .t/ C x .d/ .t/ where x .c/ and x .d/ are orthogonal processes having PSD functions
.c/ .d/ Px .c/ . f / D Px . f / and Px .d/ . f / D Px . f /
(1.223)
(1.224)
xi e j2 fi t
(1.225)
where fxi g are orthogonal random variables (r.v.s) having statistical power E[jxi j2 ] D Mi where Mi is dened in (1.220). i 2I (1.226)
48
Observation 1.5 The spectral lines of the PSD identify the periodic components in the process. Denition 1.12 A WSS random process is said to be asymptotically uncorrelated if the following two properties hold: 1/ lim rx . / D jmx j2
!1
(1.227) (1.228)
2/ cx . / D rx . /
The property 1) denotes that x.t/ and x.t For such processes, one can prove that r.c/ . / D cx . / x
(1.229)
.d/ Hence Px . f / D jmx j2 . f /, and the process exhibits at most a spectral line at the origin.
5. Px . f / D Px .
Denition 1.13 (White random process) The zero-mean random process x.t/, t 2 <, is called white if
2 rx . / D x . /
(1.232) (1.233)
In this case
2 Px . f / D x
i.e. Px is a constant.
49
(1.234) (1.235)
(1.237) (1.238)
3. Substitute the expressions of the products in the previous equations using the rule Yi Y ! P yi y j j
X` Xm
(1.239) (1.240)
! Px` xm
Then
P y1 D jH1 j2 Px1 C jH2 j2 Px2 C H1 H2 Px1 x2 C H1 H2 Px1 x2
P y2 D jH2 H3 j2 Px2
P y1 y2 D H1 H2 H3 Px1 x2 C jH2 j2 H3 Px2
h1 h2 h3
50
h x
The relation (1.245) is of particular interest since it relates the spectral density of the output process of a lter to the spectral density of the input process, through the frequency response of the lter. In the particular case in which y and z have disjoint passbands, i.e. P y . f /Pz . f / D 0, then, from (1.231), r yz . / D 0, and y ? z.
1.7.3
Let fx.k/g and fy.k/g be two discrete-time random processes. Denitions and properties of Section 1.7.1 remain valid also for discrete-time processes: the only difference is that the correlation is now dened on discrete time and is called autocorrelation sequence (ACS). It is, however, interesting to review the properties of PSDs. Given a discrete-time WSS random process x, the PSD is obtained as Px . f / D Tc F[rx .n/] D Tc
C1 X nD 1
rx .n/e
j2 f nTc
(1.247)
We note a further property: Px . f / is a periodic function of period 1=Tc . The inverse transformation yields: rx .n/ D Z
1 2Tc 1 2Tc
Px . f /e j2 f nTc d f
(1.248)
Px . f / d f
(1.249)
51
Denition 1.14 (White random process) A discrete-time random process fx.k/g is white if
2 rx .n/ D x n
(1.250)
(1.251)
Denition 1.15 If the samples of the random process fx.k/g are statistically independent and identically distributed we say that fx.k/g has i.i.d. samples.
cx .n/ e f
j2 f nTc
(1.252)
D jmx j
` Tc
(1.253)
We note that, if the process has non-zero mean value, the PSD exhibits lines at multiples of 1=Tc . Example 1.7.1 We calculate the PSD of an i.i.d. sequence fx.k/g. From ( Mx nD0 rx .n/ D 2 n 6D 0 jmx j it follows that ( cx .n/ D Then
2 r.c/ .n/ D x n x .c/ 2 Px . f / D x Tc 2 x 0
(1.254)
nD0 n 6D 0
(1.255)
(1.256) f ` Tc (1.257)
52
rx .n/z
(1.258)
From the comparison of (1.258) with (1.247), the PSD of x.k/ is related to Px .z/ by Px . f / D Tc Px .e j2 f Tc / (1.259)
Using Table 1.3, we obtain the relations between ACS and PSD listed in Table 1.6. Let the deterministic autocorrelation of h be dened as 16 rh .n/ D
C1 X kD 1
h.k/h .k
n/ D [h.m/ h . m/].n/
(1.260)
rh .n/ z
D H .z/ H
1 z
(1.261)
In case Ph .z/ is a rational function, from (1.261) one deduces that, if Ph .z/ has a pole (zero) of the type e j' jaj, it also has a corresponding pole (zero) of the type e j' =jaj. Consequently the poles (and zeros) of Ph .z/ come in pairs of the type e j' jaj, e j' =jaj. From the last relation in Table 1.6 one gets the relation between the PSDs of input and output signals, given by P y . f / D Px . f /jH .e j2 f Tc /j2 In the case of white noise input, then Py .z/ D
2 x H .z/H
(1.262)
1 z
(1.263)
Table 1.6 Relations between ACS and PSD for discrete-time processes through a linear lter.
ACS r yx .n/ D rx h.n/ rx y .n/ D [rx .m/ h . m/].n/ r y .n/ D rx y h.n/ D rx rh .n/
PSD Pyx .z/ D Px .z/H .z/ Px y .z/ D Px .z/H .1=z / Py .z/ D Px y .z/H .z/ D Px .z/H .z/H .1=z /
16 In this text we use the same symbol to indicate the correlation between random processes and the correlation
53
and
2 P y . f / D Tc x jH .e j2 f Tc /j2
(1.264)
In other words, P y . f / has the same shape as the lter frequency response. In the case of real lters 1 D H .z 1 / H z
(1.265)
Among the various applications of (1.264), it is worth mentioning the process synthesis, which deals with the generation of a random process having a pre-assigned PSD. Two methods are shown in Section 4.6.6.
where the integration is over an arbitrarily chosen interval 1=Tc . Then the function Py .z/ can be factorized as follows: 1 2 Q Q Py .z/ D f 0 F.z/ F (1.267) z where Q F.z/ D 1 C fQ1 z
1
C fQ2 z
(1.268)
is monic, minimum phase, and associated with a causal sequence f1; fQ1 ; fQ2 ; : : : g. The factor f 0 in (1.267) is the geometric mean of Py .e j2 f Tc /: Z 2 (1.269) log Py .e j2 f Tc / d f log f 0 D Tc
1=Tc
The logarithms in (1.266) and (1.269) may have any common base. The PaleyWiener criterion implies that Py .z/ may have only a discrete set of zeros Q on the unit circle, and that the spectral factorization (1.267) (with the constraint that F.z/ Q is causal, monic and minimum phase) is unique. For rational Py .z/, the function f 0 F.z/ is obtained by extracting the poles and zeros of Py .z/ that lie inside the unit circle (see Q (1.526) and the considerations relative to (1.261)). Moreover, in (1.267) f 0 F .1=z / is the Q ; fQ ; 1g, associated with poles and zeros z-transform of an anti-causal sequence f 0 f: : : ; f 2 1 of Py .z/ that lie outside the unit circle.
54
1.7.4
Denition 1.16 A WSS random process x is said to be PB (BB) if its PSD is of PB (BB) type.
For the same input x, the output of the two lters is respectively x .C/ and x . / . We nd that x.t/ D x .C/ .t/ C x . / .t/ with x . / .t/ D x .C/ .t/. The following relations are valid Px .C/ . f / D jH.C/ . f /j2 Px . f / D Px . f /1. f / Px . / . f / D jH. / . f /j2 Px . f / D Px . f /1. and Px .C/ x . / . f / D 0 as x .C/ and x .
/
(1.271)
(1.272) (1.273)
f/
(1.274)
have non-overlapping passbands. Then x .C/ ? x . / , hence (1.271) yields Px . f / D Px .C/ . f / C Px . / . f / (1.275)
rx .a/ . / D 4rx .C/ . / and Px .a/ . f / D 4Px .C/ . f / Moreover, being x .a/ D 2x . / , it follows that x .a/ ? x .a/ and rx .a/ x .a/ . / D 0 The complex envelope x .bb/ is related to x .a/ by (1.152) and rx .bb/ . / D rx .a/ . /e
j2 f 0
(1.277)
(1.278)
(1.279)
55
hence Px .bb/ . f / D Px .a/ . f C f 0 / D 4Px .C/ . f C f 0 / Moreover, from (1.278) it follows that x .bb/ ? x .bb/ . Using (1.280), (1.275) can be written as Px . f / D 1 [Px .bb/ . f 4 Finally, from
.bb/ x I .t/ D Re[x .bb/ .t/] D
(1.280)
f 0 / C Px .bb/ .
f 0 /]
(1.281)
(1.282)
and
.bb/ x Q .t/ D Im[x .bb/ .t/] D
(1.283)
rx .bb/ . / D rx .bb/ . /
Q I
Px .bb/ x .bb/ . f / D
Q I
1 [P .bb/ . f / 4j x
Q I
f /]
Q
(1.288) (1.289)
rx .bb/ x .bb/ . / D
I Q
rx .bb/ x .bb/ . / D
rx .bb/ x .bb/ . /
The second equality in (1.289) follows from Property 4 of ACS. From (1.289) we note that rx .bb/ x .bb/ . / is an odd function. Moreover, from (1.288) one
.bb/ .bb/ ? x Q only if Px .bb/ is an even function; in any case the random variables gets x I
I Q I Q
.bb/ .bb/ x I .t/ and x Q .t/ are always orthogonal since rx .bb/ x .bb/ .0/ D 0. Referring to the block
diagram in Figure 1.27b, as Px .h/ . f / D Px . f / and Px .h/ x . f / D one gets rx .h/ . / D rx . / Then rx .bb/ . / D rx .bb/ . / D rx . / cos 2 f 0 C r.h/ . / sin 2 f 0 x
I Q
j sgn. f / Px . f /
(1.290)
(1.291)
(1.292)
56
(1.293)
In terms of statistical power the following relations hold: rx .C/ .0/ D rx . / .0/ D 1 rx .0/ 2 rx .bb/ .0/ D rx .a/ .0/ D 4rx .C/ .0/ D 2rx .0/ rx .bb/ .0/ D rx .bb/ .0/ D rx .0/
I Q
rx .h/ .0/ D rx .0/ Example 1.7.2 Let x be a WSS process with power spectral density f f0 f C f0 N0 rect C rect Px . f / D 2 B B depicted in Figure 1.33. It is immediate to get Px .a/ . f / D 2N0 rect and f Px .bb/ . f / D 2N0 rect B Then Px .bb/ . f / D Px .bb/ . f / D
I Q
(1.298)
f B
f0
(1.299)
(1.300)
1 P .bb/ . f / D N0 rect 2 x
f B
(1.301)
Cyclostationary processes
In short, we have seen that, if x is a real passband WSS process, then its complex envelope is WSS, and x .bb/ ? x .bb/ . The converse is also true: if x .bb/ is a WSS process and x .bb/ ? x .bb/ , then x.t/ D Re[x .bb/ .t/ e j2 f 0 t ] is WSS with PSD given by (1.281). If x .bb/ is WSS, however, with rx .bb/ x .bb/ . / 6D 0 (1.303) (1.302)
57
N0 2 - f0 B - f0 B - f0 + 2 2 2N 0
f0 -
B 2
f0 +
B f R 2
0 2N 0 P x(bb) (f)
f0 -
B 2
f0
f0 +
B f R 2
B 2 N0
f R
B 2
B 2
f R
observing (1.302) we nd that the autocorrelation of x.t/ is a periodic function in t of period 1= f 0 : rx .t; t / D 1 [rx .bb/ . /e j2 f 0 C r .bb/ . /e 4 x C rx .bb/ x .bb/ . /e
j2 f 0 j4 f 0 t j2 f 0
C r .bb/ x .bb/ . /e j2 f 0 e x
j4 f 0 t
(1.304) ]
In other words, x is a cyclostationary process of period T0 D 1= f 0 .17 In this case it is convenient to introduce the average correlation Z T0 1 rx .t; t / dt rx . / D N T0 0
(1.305)
17 To be precise, x is cyclostationary in mean value with period T D 1= f , while it is cyclostationary in 0 0 correlation with period T0 =2.
58
whose Fourier transform is the average power spectral density 1 N N Px . f / D F[rx . /] D T0 where Px . f; t/ D F [rx .t; t /] (1.307) Z
0 T0
Px . f; t/ dt
(1.306)
In (1.307) F denotes the Fourier transform with respect to the variable . In our case, it is N Px . f / D 1 [Px .bb/ . f 4 as in the stationary case (1.281). Example 1.7.3 Let x.t/ be a modulated DSB signal (see (1.189)) x.t/ D a.t/ cos.2 f 0 t C '0 / (1.309) f 0 / C Px .bb/ . f f 0 /] (1.308)
with a.t/ real random BB WSS process with bandwidth Ba < f 0 and autocorrelation ra . /. From (1.192) it results in x .bb/ .t/ D a.t/ e j'0 . Hence we have rx .bb/ . / D ra . / rx .bb/ x .bb/ . / D ra . / e j2'0 (1.310)
Because ra . / is not identical to zero, observing (1.303) we nd that x.t/ is cyclostationary with period 1= f 0 . From (1.308) the average PSD of x is given by N Px . f / D 1 [Pa . f 4 f 0 / C Pa . f C f 0 /] (1.311)
Ma
(1.312)
We note that one nds the same result (1.311) assuming that '0 is a uniform r.v. in [0; 2 /; in this case x turns out to be WSS. Example 1.7.4 Let x.t/ be a modulated single sideband (SSB) with an upper sideband, x.t/ D Re[ 1 .a.t/ C ja .h/ .t//e j .2 f 0 tC'0 / ] 2 D
1 2
1 2
(1.313)
where a .h/ .t/ is the Hilbert transform of a.t/, a real WSS random process with autocorrelation ra . / and bandwidth Ba .
59
We note that the modulating signal .a.t/ C ja .h/ .t// coincides with the analytic signal a .a/ .t/ and it has spectral support only for positive frequencies, therefore it is one half of a.t/. Being x .bb/ .t/ D 1 .a.t/ C ja .h/ .t//e j'0 2 it results that x .bb/ and x .bb/ have non-overlapping passbands and rx .bb/ x .bb/ . / D 0 The process (1.313) is then stationary with Px . f / D 1 [Pa .C/ . f 4 f 0 / C Pa .C/ . f f 0 /] (1.315) (1.314)
where a .C/ .t/ is dened in (1.271). In this case x has bandwidth equal to Ba and statistical power given by Mx D
1 4
Ma
(1.316)
Example 1.7.5 (DSB and SSB demodulators) Let the signal r.t/ be the sum of a desired part x.t/ and additive white noise w.t/ with PSD equal to Pw . f / D N0 =2, r.t/ D x.t/ C w.t/ (1.317)
where the signal x is modulated DSB (1.309). To obtain the signal a.t/ from r.t/, one can use the coherent demodulation scheme illustrated in Figure 1.34 (see Figure 1.30b) where h is an ideal lowpass lter, having a frequency response f H. f / D H0 rect (1.318) 2Ba Let ro be the output signal of the demodulator, given by the sum of the desired part xo and noise wo : ro .t/ D xo .t/ C wo .t/ (1.319)
60
We evaluate now the ratio between the powers of the signals in (1.319), 3o D in terms of the reference ratio 0D Mx .N0 =2/ 2Ba (1.321) Mxo M wo (1.320)
Using the equivalent block scheme of Figure 1.34 and (1.192), we have r .bb/ .t/ D a.t/ e j'0 C w.bb/ .t/ with Pw.bb/ . f / D 2N0 1. f C f 0 /. Being h a.t/ D H0 a.t/ it results xo .t/ D Re[h./ D Hence we get Mxo D
2 H0 Ma cos2 .'0 4 1 2
(1.322)
(1.323)
j'1
H0 a.t/ cos.'0 2
'1 /
(1.325)
In the same baseband equivalent scheme, we consider the noise weq at the output of lter h; we nd Pweq . f / D D
1 4
2 H0 f N0 rect 2 2Ba
Being now w WSS, w.bb/ is uncorrelated with w.bb/ and thus weq with weq . Then, from
(1.327)
2 H0 f N0 rect 4 2Ba
(1.328)
(1.329)
61
'1 /
N0 2Ba
D 0 cos2 .'0
'1 /
(1.330)
It is interesting to observe that, at the demodulator input, the ratio between the power of the desired signal and the power of the noise in the passband of x is given by 3i D For '1 D '0 then 3o D 23i (1.333) Mx 0 D .N0 =2/ 4Ba 2 (1.332)
In other words, measuring the noise power in a passband equal to that of the desired signal, the DSB demodulator yields a gain of 2 in signal-to-noise ratio. We will now analyze the case of a SSB signal x.t/ (see (1.313)), coherently demodulated, following the scheme of Figure 1.35, where h P B is a lter used to eliminate the noise that otherwise, after the mixer, would have fallen within the passband of the desired signal. The ideal frequency response of h P B is given by f f 0 Ba =2 f f 0 Ba =2 C rect (1.334) H P B . f / D rect Ba Ba Note that in this scheme we have assumed the phase of the receiver carrier equal to that of the transmitter, to avoid distortion of the desired signal. Being f Ba =2 H.bb/ . f / D 2 rect (1.335) PB Ba the lter of the baseband-equivalent scheme is given by h eq .t/ D
1 2
h .bb/ h.t/ PB
(1.336)
Figure 1.35. (a) Coherent SSB demodulator and (b) baseband-equivalent scheme.
62
Ba =2 Ba
(1.337)
We now evaluate the desired component xo .t/. Using the fact x .bb/ h eq .t/ D H0 x .bb/ .t/, it results in xo .t/ D Re[h eq ./ 1 e j'0 1 .a./ C j a .h/ .// e j'0 ].t/ 2 2 D D H0 Re[a.t/ C j a .h/ .t/] 4 (1.338)
H0 a.t/ 4 In the baseband-equivalent scheme, the noise weq at the output of h eq has a PSD given by Pweq . f / D jHeq . f /j2 2N0 1. f C f 0 / (1.339) f Ba =2 N0 2 H0 rect D 2 Ba and using (1.285), which is valid because weq ? weq ,
1 4
From the relation wo D weq;I we have H2 1 f (1.340) Pwo . f / D [Pweq . f / C Pweq . f /] D 0 N0 rect 4 8 2Ba and H2 Mwo D 0 N0 2Ba (1.341) 8 Then we obtain 2 .H0 =16/ Ma 3o D (1.342) 2 .H0 =8/ N0 2Ba which using (1.316) and (1.321) can be written as 3o D 0 (1.343) We note that the SSB system yields the same performance (for '1 D '0 ) as a DSB system, even though half of the bandwidth is required. Finally, it results in Mx D 3o (1.344) 3i D .N0 =2/ 2Ba
Observation 1.6 We note that also for the simple examples considered in this section the desired signal is analyzed via the various transformations, whereas the noise is analyzed via the PSD. As a matter of fact, we are typically interested only in the statistical power of the noise at the system output. The demodulated signal xo .t/, on the other hand, must be expressed as the sum of a desired component proportional to a.t/ and an orthogonal component that represents the distortion, which is, typically, small and has the same effects as noise. In the previous example, the considered systems do not introduce any distortion since xo .t/ is proportional to a.t/.
63
1.8
Denition
Given the discrete-time wide-sense stationary random process fx.k/g, we introduce the random vector with N components xT .k/ D [x.k/; x.k The N N autocorrelation matrix of 2 rx .0/ 6 rx .1/ 6 R D E[x .k/xT .k/] D 6 : : 4 : rx .N x .k/ 1/; : : : ; x.k is given by rx . 1/ rx .0/ : : : rx . N C 1/ rx . N C 2/ :: : 2/ rx .0/ N C 1/] 3 7 7 7 5 (1.346) (1.345)
1/ rx .N
Properties
1. R is Hermitian: R H D R. For real random processes R is symmetric: RT D R. 2. R is a Toeplitz matrix, i.e. all elements along any diagonal are equal. 3. R is positive semi-denite and almost always positive denite. Indeed, taking an arbitrary vector vT D [v0 ; : : : ; v N 1 ], and letting y D xT .k/v, we have E[jyj2 ] D E[v H x .k/xT .k/v] D v H Rv D
N 1N 1 XX i D0 jD0
vi rx .i
j/v j 0
(1.347)
If v H Rv > 0, 8v, then R is said to be positive denite and all its principal minor determinants are positive; in particular R is non-singular.
Eigenvalues
We indicate by det[R] the determinant of a matrix R. The eigenvalues of R are the solutions i , i D 1; : : : ; N , of the characteristic equation of order N det[R I] D 0 (1.348) (1.349) and the corresponding column eigenvectors ui satisfy the equation Rui D i ui
Example 1.8.1 Let fw.k/g be a white noise process. Its autocorrelation matrix R assumes the form 3 2 2 w 0 0 2 6 0 w 0 7 7 6 RD6 : : : (1.350) : 7 4 : : :: : 5 : : : 2 0 0 w
64
(1.351)
and ui can be any arbitrary vector Example 1.8.2 We dene a complex-valued sinusoid as x.k/ D e j .!kC'/ 2 6 6 6 RD6 6 4 ! D 2 f Tc 3 7 7 7 7 7 5 (1.354) (1.353) 1i N (1.352)
e : : : e j .N
j!
e e :: :
j .N 1/! j .N 2/!
: : : 1
2/!
One can see that the rank of R is 1 and it will therefore have only one eigenvalue. A possible solution is given by 1 D N and the relative eigenvector is
T u1 D [1; e j! ; : : : ; e j .N 1/!
(1.355)
(1.356)
Other properties
1. From Rm u D m u we obtain the relations of Table 1.7. 2. If the eigenvalues are distinct, then the eigenvectors are linearly independent:
N X i D1
ci ui 6D 0
(1.357)
for all combinations of fci g, i D 1; 2; : : : ; N , not all equal to zero. Therefore, in this case, the eigenvectors form a basis in < N .
Table 1.7 Correspondence between eigenvalues and eigenvectors of four matrices.
R Eigenvalue Eigenvector i ui
Rm
m i ui
I .1
R i / ui
i 1 ui
65
3. The trace of a matrix R is dened as the sum of the elements of the main diagonal, and we indicate it with tr[R]. It holds tr R D
N X i D1
(1.358)
The ratio (1.360) is dened as Rayleigh quotient. As R is positive semi-denite, uiH Rui 0, from which i 0. 2. If the eigenvalues of R are distinct, then the eigenvectors are orthogonal. In fact, from (1.349) one gets: uiH Ru j D j uiH u j uiH Ru j D i uiH u j Subtracting the second equation from the rst: 0 D . j and since j i /uiH u j (1.363) (1.361) (1.362)
3. If the eigenvalues of R are distinct and their corresponding eigenvectors are normalized, i.e. ( 1 iD j 2 H jjui jj D ui ui D (1.364) 0 i 6D j then the matrix U, whose columns are the eigenvectors of R, U D [u1 ; u2 ; : : : ; u N ] (1.365)
66
D UH
(1.366)
This property is an immediate consequence of the orthogonality of the eigenvectors fui g. Moreover, if we dene the matrix as 2 3 1 0 0 6 0 0 7 2 6 7 7 (1.367) D6 : : : : 7 6 : : : : : 5 4 : : : 0 0 N we get U H RU D From (1.368) we obtain the following important relations: R D U UH D and I R D U.I /U H D
N X i D1 N X i D1
(1.368)
i ui uiH
(1.369)
.1
i /ui uiH
(1.370)
4. The eigenvalues of a positive semi-denite autocorrelation matrix R and the PSD of x are related by the inequalities, minfPx . f /g i maxfPx . f /g
f f
i D 1; : : : ; N
(1.371)
In fact, let Ui . f / be the Fourier transform of the sequence represented by the elements of ui : Ui . f / D
N X nD1
u i;n e
j2 f nTc
(1.372)
where u i;n is the n-th element of the eigenvector ui . Observing that uiH Rui D
N N XX nD1 mD1 u i;n rx .n
m/u i;m
(1.373)
and using (1.248) and (1.372), the preceding equation can be written as uiH Rui D Z
1 2Tc 1 2Tc
Px . f /
N X nD1
u i;n e j2 f nTc
N X mD1
u i;m e
j2 f mTc
df (1.374)
Z D
1 2Tc 1 2Tc
Px . f / jUi . f /j2 d f
67
Px . f / jUi . f /j2 d f
1 2Tc 1 2Tc
from which (1.371) follows. If we indicate with min and max , respectively, the minimum and maximum eigenvalue of R, in view of the latter point we can dene the eigenvalue spread as: .R/ D max f fPx . f /g max min min f fPx . f /g (1.376)
From (1.376) we observe that .R/ may assume large values in the case Px . f / exhibits large variations. Moreover, .R/ assumes the minimum value of 1 for a white process.
1.9
Before reviewing some important random processes, we recall the denition of Gaussian complex-valued random vector. Example 1.9.1 A r.v. with a Gaussian distribution can be generated from two r.v.s with uniform distribution (see Appendix 1.B for an illustration of the method). Example 1.9.2 Let xT D [x1 ; : : : ; x N ] be a real Gaussian random vector, xi 2 N mi ; i2 . The joint probability density function is px . / D [.2 / N det C N ]
1 1 1 T 2 e 2 . mx / C N . mx /
(1.377)
where T D [1 ; : : : ; N ], mx D E[x] is the vector of mean values and C N D E[.x mx /.x mx /T ] is the covariance matrix. Example 1.9.3 Let xT D [x1;I C j x1;Q ; : : : ; x N ;I C j x N ;Q ] be a complex-valued Gaussian random vector. If the in-phase component xi;I and the quadrature component xi;Q are uncorrelated, E[.xi;I and, moreover,
2 2 2 xi;I D xi;Q D 1 xi 2
mxi;I /.xi;Q
mxi;Q /] D 0
i D 1; 2; : : : ; N
(1.378)
(1.379)
68
. mx / H C N 1 . mx /
(1.380)
with the vector of mean values and the covariance matrix given by mx D E[x] D E[x I ] C j E[x Q ] C N D E[.x mx /.x mx / ]
H
(1.381) (1.382)
The vector x is called a circularly symmetric Gaussian random vector. Example 1.9.4 Let xT D [x1 .t1 /; : : : ; x N .t N /] be a complex-valued Gaussian (vector) process, with each element xi .ti / having real and imaginary components that are uncorrelated Gaussian r.v.s with zero mean and equal variance for all values of ti . The vector x is called a circularly symmetric Gaussian random process. The joint probability density function in this case results in px . / D [ N det C]
1
HC
(1.383)
where C is the covariance matrix of [x1 .t1 /; x2 .t2 /; : : : ; x N .t N /]. Example 1.9.5 Let x.t/ D A sin.2 f t C'/ be a real-valued sinusoidal signal with ' r.v. uniform in [0; 2 /, for which we will use the notation ' 2 U[0; 2 /. The mean of x is mx .t/ D E[x.t/] Z 2 1 A sin.2 f t C a/ da D 2 0 D0 and the autocorrelation function is given by Z 2 1 rx . / D A sin.2 f t C a/A sin[2 f .t 2 0 A2 cos.2 f / D 2 Example 1.9.6 Given N real-valued sinusoidal signals x.t/ D
N X i D1
(1.384)
/ C a] da (1.385)
Ai sin.2 f i t C 'i /
(1.386)
69
with f'i g statistically independent uniform r.v.s in [0; 2 /, from Example 1.9.5 it is able to obtain the mean value mx .t/ D and the autocorrelation function rx . / D
N X A2 i cos.2 f i / 2 i D1 N X i D1
mxi .t/ D 0
(1.387)
(1.388)
We note that, according to the Denition 1.12, page 48, the process (1.386) is not asymptotically uncorrelated. Example 1.9.7 Given N complex-valued sinusoidal signals x.t/ D
N X i D1
Ai e j .2 fi tC'i /
(1.389)
with f'i g statistically independent uniform r.v.s in [0; 2 /, following a similar procedure to that used in Examples 1.9.5 and 1.9.6, we nd rx . / D
N X i D1
jAi j2 e j2 fi
(1.390)
We note that the process (1.390) is not asymptotically uncorrelated. Example 1.9.8 Let the discrete-time random process y.k/ D x.k/ C w.k/ be given by the sum of the 2 random process x.k/ of Example 1.9.7 and white noise w.k/ with variance w . Moreover, we assume fx.k/g and fw.k/g are uncorrelated. In this case r y .n/ D
N X i D1 2 jAi j2 e j2 fi nTc C w n
(1.391)
Example 1.9.9 We consider a signal obtained by pulse-amplitude modulation (PAM), expressed as y.t/ D
C1 X kD 1
x.k/ h T x .t
kT /
(1.392)
70
x(k) T
hTx
y(t)
The signal y.t/ is the output of the system shown in Figure 1.36, where h T x is a niteenergy pulse, and fx.k/g is a discrete-time (with T -spaced samples) WSS sequence, having power spectral density Px . We note that Px . f / is a periodic function of period 1=T . Let rh T x . / be the deterministic autocorrelation of the signal h T x : Z C1 h T x .t/h x .t / dt D [h T x .t/ h x . t/]. / (1.393) rh T x . / D T T
1
with Fourier transform equal to jHT x . f /j2 . In general y is a cyclostationary process of period T . In fact we have 1. Mean m y .t/ D mx 2. Correlation r y .t; t / D
C1 X iD 1 C1 X kD 1
h T x .t
kT /
(1.394)
rx .i/
C1 X mD 1
h T x .t
.i C m/ T /h x .t T
mT /
(1.395)
If we introduce the average spectral analysis Z 1 T my D m y .t/ dt D mx HT x .0/ T 0 Z C1 1 T 1 X r y .t; t / dt D rx .i/rh T x . r y . / D N T 0 T iD 1 and 2 1 N P y . f / D F[r y . /] D HT x . f / Px . f / N T
(1.398)
we observe that the modulator of a PAM system may be regarded as an interpolator lter with frequency response HT x =T . 3. Average power for a white noise input For a white noise input with power Mx , from (1.397) the average statistical power of the output signal is given by N M y D Mx where E h D R C1
1
Eh T
(1.399)
71
4. Moments of y for a circularly symmetric i.i.d. input Let x.k/ be a complex-valued random circularly symmetric sequence with zero mean (see (1.378) and (1.379)), i.e. letting x I .k/ D Re[x.k/] we have
2 2 E[x I .k/] D E[x Q .k/] D
(1.400)
(1.401)
and E[x I .k/ x Q .k/] D 0 These two relations can be merged into one,
2 E[x 2 .k/] D E[x I .k/] 2 E[x Q .k/] C 2 j E[x I .k/ x Q .k/] D 0
(1.402)
(1.403)
Filtering the i.i.d. input signal fx.k/g by using the scheme depicted in Figure 1.36, and observing the relation r yy .t; t then rx x .i/ D E[x 2 .k/].i/ D 0 and r yy .t; t / D 0 (1.406) (1.405) / D
C1 X iD 1
rx x .i/
C1 X mD 1
h T x .t
.i C m/T /h T x .t
mT / (1.404)
that is y.t/ ? y .t/. In particular we nd that y.t/ is circularly symmetric, i.e. E[y 2 .t/] D 0 (1.407)
We note that the condition (1.406) can be obtained assuming the less stringent condition that x ? x ; on the other hand, this requires that the following two conditions are veried rx I .i/ D rx Q .i/ and rx I x Q .i/ D rx I x Q . i/ (1.409) (1.408)
Observation 1.7 It can be shown that if the lter h T x has a bandwidth smaller than 1=.2T / and x is a WSS sequence, then y is WSS with spectral density given by (1.398).
72
Example 1.9.10 Let us consider a PAM signal sampled with period TQ D T =Q 0 , where Q 0 is a positive integer number. Let h p D h T x . p TQ / (1.410) yq D y.q TQ / from (1.392) it follows C1 X yq D x.k/ h q k Q 0 (1.411)
kD 1
If Q 0 6D 1, (1.411) describes the inputoutput relation of an interpolator lter (see (1.609)). We recall the statistical analysis given in Table 1.6, page 52. We denote with H. f / the Fourier transform (see (1.84)) and with rh .n/ the deterministic autocorrelation (see (1.260)) of the sequence fh p g. We also assume that fx.k/g is a WSS random sequence with mean mx and autocorrelation rx .n/. In general, fyq g is a cyclostationary random sequence of period Q 0 with 1. Mean m y .q/ D mx 2. Correlation r y .q; q n/ D
C1 X iD 1 C1 X kD 1 C1 X mD 1
hq
k Q0
(1.412)
rx .i/
hq
.i Cm/Q 0
hq
n m Q0
(1.413)
By the average spectral analysis we obtain X H.0/ 1 Q0 1 my D N m y .q/ D mx Q 0 qD0 Q0 where H.0/ D and r y .n/ D N X 1 Q0 1 r y .q; q Q 0 qD0 n/ D
C1 1 X rx .i/ rh .n Q0 i D 1 C1 X pD 1
(1.414)
hp
(1.415)
i Q0/
(1.416)
Consequently, the average PSD is given by 2 N y . f / D TQ F[r y .n/] D 1 H. f / Px . f / P N Q 0 If fx.k/g is white noise with power Mx , from (1.416) it results in rh .n/ Q0 In particular the average power of the lter output signal is given by r y .n/ D Mx N N M y D Mx Eh Q0 (1.418) (1.417)
(1.419)
73
P N where E h D C1 1 jh p j2 is the energy of fh p g. We point out that the condition M y D Mx pD is satised if the energy of the lter impulse response is equal to the interpolation factor Q 0 .
1.10
Matched lter
Referring to Figure 1.37, we consider a nite-energy signal pulse g in the presence of additive noise w having zero mean and power spectral density Pw . The signal x.t/ D g.t/ C w.t/ (1.420) is ltered with a lter having impulse response g M . We indicate with gu and wu respectively the desired signal and the noise component at the output: gu .t/ D g M g.t/ wu .t/ D g M w.t/ The output is expressed as y.t/ D gu .t/ C wu .t/ (1.423) (1.421) (1.422)
We now suppose that y is observed at a given instant t0 . The problem is to determine g M so that the ratio between the squared amplitude gu .t0 / and the power of the noise component wu .t0 / is maximum, i.e. g M : max
gM
(1.424)
(1.425)
where K is a constant. In other words, the best lter selects the frequency components of the desired input signal and weights them with weights that are inversely proportional to the noise level. Proof. gu .t0 / coincides with the inverse Fourier transform of G M . f /G. f / evaluated in t D t0 , while the power of wu .t0 / is equal to Z C1 (1.426) Pw . f /jG M . f /j2 d f rwu .0/ D
1
y(t)
t0
e -j2 ft0
74
G M . f /G. f /e
j2 f t0
1 C1 1
2 df
C1 1
2 G. f / j2 f t0 G M . f / Pw . f / p df e Pw . f / Z C1 Pw . f /jG M . f /j2 d f
1
p where the integrand at the numerator was divided and multiplied by Pw . f /. Implicitly it is assumed that Pw . f / 6D 0. Applying the Schwarz inequality (see Section 1.1) to the functions p (1.428) G M . f / Pw . f / and G. f / e p Pw . f / it turns out jgu .t0 /j2 rwu .0/ Z
C1 1 j2 f t0
(1.429)
2 Z pG. f / e j2 f t0 d f D P .f/
w
C1 1
2 pG. f / d f P . f /
w
(1.430)
Therefore the maximum value is equal to the right-hand side of (1.430) and is achieved for p G . f / e G M . f / Pw . f / D K p Pw . f /
j2 f t0
(1.431)
(1.432)
from which comes the name of matched lter (MF), i.e. matched to the input signal pulse. The desired signal pulse at the lter output has the frequency response Gu . f / D K jG. f /j2 e
j2 f t0
(1.434)
75
x(t)=g(t)+w(t) gM
y(t)=Krg (t - t0 ) + wu (t)
t0
y(t0 )
gM (t)=Kg*(t0 -t)
Figure 1.38. Matched lter for an input pulse in the presence of white noise.
(1.435)
If E g is the energy of g, using the relation E g D rg .0/ the maximum of the functional (1.424) becomes jK j2 r2 .0/ Eg jgu .t0 /j2 g D D 2 r .0/ rwu .0/ Pw Pw jK j g (1.437)
In Figure 1.39 the different pulse shapes are illustrated for a signal pulse g with limited duration tg . Note that in this case the matched lter has also limited duration and it is causal if t0 tg . Example 1.10.1 (MF for a rectangular pulse) Let g.t/ D wT .t/ D rect with rg . / D T 1 j j rect T 2T (1.439) t T =2 T (1.438)
For t0 D T , the matched lter is proportional to g g M .t/ D K wT .t/ and the output pulse in the absence of noise is equal to 8 > < K T 1 t T 0 < t < 2T T gu .t/ D > :0 elsewhere (1.440)
(1.441)
76
g(t)
0 t0 = 0 gM (t)
tg
-tg t0 = t g
0 gM (t)
0 r g (t)
tg
-tg
tg
1.11
The functions that have been introduced in the previous sections for the analysis of random processes give a valid statistical description of an ensemble of realizations of a random process. We investigate now the possibility of moving from ensemble averaging to time averaging, that is we consider the problem of estimating a statistical descriptor of a random process from the observation of a single realization. Let x be a discrete-time WSS random process having mean mx . If in the limit it holds18 X 1 K 1 lim x.k/ D E[x.k/] D mx (1.442) K !1 K kD0
1 P K 1 x.k/ K kD0
18 The limit is meant in the mean-square sense, that is the variance of the r.v.
mx
vanishes
for K ! 1.
77
then x is said to be ergodic in the mean. In other words, for a process for which the above limit is true, the time-average of samples tends to the statistical mean as the number of samples increases. We note that the existence of the limit (1.442) implies the condition 2 2 3 1 K 1 X lim E 4 (1.443) x.k/ mx 5 D 0 K kD0 K !1 or equivalently 1 lim K !1 K
K 1 X nD .K 1/
jnj cx .n/ D 0 K
(1.444)
From (1.444) we see that for a random process to be ergodic in the mean, some conditions on the second-order statistics must be veried. Analogously to denition (1.442), we say that x is ergodic in correlation if in the limits it holds: lim X 1 K 1 x.k/x .k K kD0 n/ D E[x.k/x .k n/] D rx .n/ (1.445)
K !1
Also for processes that are ergodic in correlation one could get a condition of ergodicity similar to that expressed by the limit (1.444). Let y.k/ D x.k/x .k n/. Observing (1.445) and (1.442), we nd that the ergodicity in correlation of the process x is equivalent to the ergodicity in the mean of the process y. Therefore it is easy to deduce that the condition (1.444) for y translates into a condition on the statistical moments of the fourth order for x. In practice, we will assume all stationary processes to be ergodic; ergodicity is, however, difcult to prove for non-Gaussian random processes. We will not consider particular processes that are not ergodic such as x.k/ D A, where A is a random variable, or x.k/ equal to the sum of sinusoidal signals (see (1.386)). The property of ergodicity assumes a fundamental importance if we observe that from a single realization it is possible to obtain an estimate of the autocorrelation function and from this, the power spectral density. Alternatively, one could prove that under the hypothesis19
C1 X nD 1
(1.446)
1 lim E 4 K !1 K Tc
K 1 X x.k/ e Tc kD0
(1.447)
Then, exploiting the ergodicity of a WSS random process, one obtains the relations among the process itself, its autocorrelation function and power spectral density shown
19 We note that for random processes with non-zero mean and/or sinusoidal components this property is not
veried. Therefore it is usually recommended that the deterministic components of the process be removed before the spectral estimation is performed.
78
Figure 1.40. Relation between ergodic processes and their statistical description.
in Figure 1.40. We note how the direct computation of the PSD, given by (1.447), makes use of a statistical ensemble of the Fourier transform of process x, while the indirect method via ACS makes use of a single realization. If we let Q X K Tc . f / D Tc F[x.k/ w K .k/] (1.448)
where w K is the rectangular window of length K (see (1.474)) and Td D K Tc , (1.447) becomes Px . f / D lim
Td !1
Q E[jXTd . f /j2 ] Td
(1.449)
The relation (1.449) holds also for continuous-time ergodic random processes, where Q XTd . f / denotes the Fourier transform of the windowed realization of the process, with a rectangular window of duration Td .
1.11.1
Given the random process fx.k/g, we wish to estimate the mean value of a related process fy.k/g: for example, to estimate the statistical power of x we set y.k/ D jx.k/j2 , while for the estimation of the correlation of x with lag n we set y.k/ D x.k/x .k n/. Based on a realization of fy.k/g, from (1.442) an estimate of the mean value of y is given by the expression my D O X 1 K 1 y.k/ K kD0 (1.450)
79
In fact, (1.450) attempts to determine the average component of the signal fy.k/g. As illustrated in Figure 1.41a, in general we can think of extracting the average component of fy.k/g using an LPF lter h having unit gain, i.e. H.0/ D 1, and suitable bandwidth B. Let K be the length of the impulse response with support from k D 0 to k D K 1. Note that for a unit step input signal the transient part of the output signal will last K 1 time instants. Therefore we assume m y D z.k/ D h y.k/ O for k K 1 (1.451)
We now compute mean and variance of the estimate. From (1.451), the mean value is given by E[m y ] D m y H.0/ D m y O (1.452)
0.035
0.03
0.025
0.02
h(k)
0.015
0.01
0.005
10
15
20 k
25
30
35
40
(a)
1.2
(b)
0.8
0.6
|H(f)|
0.4
0.2
0
0.2
0.05
0.1
0.15
0.2
0.25 fT
c
0.3
0.35
0.4
0.45
0.5
(c)
Figure 1.41. (a) Time average as output of a narrow band lowpass lter. (b) Typical impulse responses: exponential lter with parameter a D 1 2 5 and rectangular window with K D 33. (c) Corresponding frequency responses.
80
as H.0/ D 1. Using the expression in Table 1.6 of the correlation of a lter output signal given the input, the variance of the estimate is given by
2 var[m y ] D y D O C1 X nD 1
rh . n/c y .n/
(1.453)
Assuming SD
C1 X nD 1 2 jc y .n/j D y C1 X jc y .n/j <1 2 y nD 1
(1.454)
and being jrh .n/j rh .0/, the variance in (1.453) is bounded by var[m y ] E h S O where E h D rh .0/. For an ideal lowpass lter f H. f / D rect 2B jfj < 1 2Tc (1.456) (1.455)
assuming as lter length K that of the principal lobe of fh.k/g, and neglecting a delay factor, it results in E h D 2B and K ' 1=B. Introducing the criterion that for a good estimate it must be var[m y ] " O with " jm y j2 , from (1.455) it follows B and K 2S " (1.459) " 2S (1.458) (1.457)
In other words, from (1.454) and (1.459), for a xed ", the length K of the lter impulse response must be larger, or equivalently the bandwidth B must be smaller, to obtain estimates for those processes fy.k/g that exhibit larger variance and/or larger correlation among samples. Because of their simple implementation, two commonly used lters are the rectangular window and the exponential lter, whose impulse responses are shown in Figure 1.41.
Rectangular window
8 < 1 h.k/ D K :0 k D 0; 1; : : : ; K elsewhere 1
(1.460)
81
sin. f K Tc / sin. f Tc /
(1.461)
For the rectangular window we have E h D 1=K and, adopting as bandwidth the frequency of the rst zero of jH. f /j, B D 1=.K Tc /. The lter output is given by z.k/ D that can be expressed as z.k/ D z.k 1/ C y.k/ y.k K K/ (1.463)
K 1 X nD0
1 y.k K
n/
(1.462)
Exponential lter
( h.k/ D .1 0 a/a k k0 elsewhere (1.464)
Moreover, E h D .1 a/=.1 C a/ and, adopting as length of h the time constant of the lter, i.e. the interval it takes for the amplitude of the impulse response to decrease of a factor e, K 1D 1 1 ' ln 1=a 1 a (1.466)
where the approximation holds for a ' 1. The 3 dB lter bandwidth is equal to BD 1 a 1 2 Tc for a > 0:9 (1.467)
The lter output has a simple expression given by the recursive equation z.k/ D az.k We note that choosing a as aD1 the expression (1.468) becomes z.k/ D z.k 1/ C 2 l .y.k/ z.k 1// (1.470) 2
l
1/ C .1
a/ y.k/
(1.468)
(1.469)
whose computation requires only two additions and one shift of l bits. Moreover, from (1.466), the lter time constant is given by K 1 D 2l (1.471)
82
General window
In addition to the two lters described above, a general window can be dened as h.k/ D Aw.k/ window20 (1.472) with fw.k/g of length K . The factor A in (1.472) is introduced to normalize the area of h to 1. We note that, for random processes with slowly time-varying statistics, the equations (1.463) and (1.470) give an expression to update the estimates.
1.11.2
Correlation estimators
1, be a realization of a random process with K samples. We
Unbiased estimate
rx .n/ D O 1 K n
K 1 X kDn
x.k/x .k
n/
n D 0; 1; : : : ; K
(1.478)
20
We dene the continuous-time rectangular window with duration Td as ( 1 0 < t < Td t Td =2 wTd .t/ D rect D Td 0 elsewhere Commonly used discrete-time windows are: 1. Rectangular window ( w.k/ D w D .k/ D 1 0 k D 0; 1; : : : ; D elsewhere 1
(1.473)
(1.474)
where D denotes the length of the rectangular window expressed in number of samples. 2. Raised cosine or Hamming window 8 0 D 11 > > k > < B 2 C 0:54 C 0:46 cos @2 A w.k/ D D 1 > > > : 0 3. Hann window 8 > > > < > > > : 0 B 0:50 C 0:50 cos @2 0 8 > > > < > > > : D 1 k 2 2 D 1 11 C A
k D 0; 1; : : : ; D elsewhere
(1.475)
k D
D 2 1
w.k/ D
k D 0; 1; : : : ; D elsewhere
(1.476)
w.k/ D
1 0
(1.477)
83
E[x.k/x .k
n/] D rx .n/
(1.479)
If the process is Gaussian, one can show that the variance of the estimate is approximately given by var[rx .n/] ' O from which it follows var[rx .n/] O !0
K !1
K .K n/2
C1 X mD 1
n/]
(1.480)
(1.481)
The above limit holds for n K . Note that the variance of the estimate increases with the correlation lag n.
Biased estimate
rx .n/ D L X 1 K 1 x.k/x .k K kDn n/ D 1 jnj rx .n/ O K (1.482)
The mean value of the biased estimate satises the following relations: jnj E[rx .n/] D 1 L rx .n/ ! rx .n/ K K !1
(1.483)
Unlike the unbiased estimate, the mean of the biased estimate is not equal to the autocorrelation function, but approaches it as K increases. Note that the biased estimate differs from the autocorrelation function by one additive constant, denoted as BIAS : BIAS D E[rx .n/] L rx .n/ (1.484)
For a Gaussian process, the variance of the biased estimate is expressed as C1 K jnj 2 1 X var[rx .n/] D L var[rx .n/] ' O [r2 .m/ C rx .m C n/rx .m K K mD 1 x
n/]
(1.485) In general, the biased estimate of the ACS exhibits a mean-square error21 larger than the unbiased, especially for large values of n. It should also be noted that the estimate does not necessarily yield sequences that satisfy the properties of autocorrelation functions: for example, the following property may not be veried: O rx .0/ jrx .n/j O n 6D 0 (1.487)
21 For example, for the estimator (1.478) the mean-square error is dened as
E[jrx .n/ O
(1.486)
84
1.11.3
After examining ACS estimators, we review some spectral density estimation methods.
1;
(1.488) Q jX . f /j2 d f
1 D K Tc
using the properties of the Fourier transform (Parseval theorem). Based on (1.488), a PSD estimator called a periodogram is given by PPER . f / D We can write (1.489) as PPER . f / D Tc and, consequently, E[PPER . f /] D Tc
K 1 X nD .K 1/ K 1 X nD .K 1/ K 1 X nD .K 1/
1 Q jX . f /j2 K Tc
(1.489)
rx .n/ e L
j2 f nTc
(1.490)
E[rx .n/]e L 1
j2 f nTc
D Tc
jnj rx .n/e K
j2 f nTc
(1.491)
D Tc W B Px . f / where W B . f / is the Fourier transform of the Bartlett window ( jnj jnj K 1 1 w B .n/ D K 0 jnj > K 1 and 1 WB . f / D K sin. f K Tc / sin. f Tc / 2 (1.493)
(1.492)
We note the periodogram estimate is affected by BIAS for nite K . Moreover, it also L exhibits a large variance, as PPER . f / is computed using the samples of rx .n/ even for lags up to K 1, whose variance is very large.
85
Welch periodogram
This method is based on applying (1.447) for nite K . Given a sequence of K samples, different subsequences of consecutive D samples are extracted. Subsequences may partially overlap. Let x .s/ be the s-th subsequence, characterized by S samples in common with the preceding subsequence x .s 1/ and with the following one x .sC1/ . In general, 0 S D=2, with the choice S D 0 yielding subsequences with no overlap and therefore with less correlation. The number of subsequences Ns is22 K D (1.494) C1 Ns D D S Let w be a window (see footnote 20 on page 82) of D samples: then x .s/ .k/ D w.k/ x.k C s.D S// k D 0; 1; : : : ; D 1 s D 0; 1; : : : ; N s 1 (1.495)
x .s/ .k/e
j2 f kTc
(1.496)
2 1 Q .s/ X . f / DTc Mw
(1.497)
is the normalized energy of the window. As a last step, for each frequency, average the periodograms: PWE . f / D X 1 Ns 1 .s/ P .f/ Ns sD0 PER (1.499)
Mean and variance of the estimate are given by E[PWE . f /] D Tc [jW./j2 Px ./]. f / where W. f / D
D 1 X kD0
(1.500)
w.k/e
j2 f kTc
(1.501)
22 The symbol bac denotes the function oor, that is the largest integer smaller than or equal to a. The symbol
dae denotes the function ceiling, that is the smallest integer larger than or equal to a.
86
Assuming the process is Gaussian and the different subsequences are statistically independent, we get23 var[PWE . f /] / 1 2 P .f/ Ns x (1.502)
Note that the partial overlap introduces correlation between subsequences. From (1.502), we see that the variance of the estimate is reduced by increasing the number of subsequences. In general, D must be large enough so that the generic subsequence represents the process and also Ns must be large to obtain a reliable estimate (see (1.502)); therefore the application of the Welch method requires many samples.
w.n/rx .n/ e O
j2 f nTc
(1.503)
where w is a window24 of length 2L C 1, with w.0/ D 1. If K is the number of samples of the realization sequence, we require that L K =5 to reduce the variance of the estimate. Then if the Bartlett window (1.493) is chosen, one nds that PBT . f / 0. In terms of the mean value of the estimate, we nd E[PBT . f /] D Tc W. f / Px . f / (1.504)
For a Gaussian process, if the Bartlett window is chosen, the variance of the estimate is given by var[PBT . f /] D 1 2 2L 2 Px . f /E w D P .f/ K 3K x (1.505)
23 The symbol / indicates proportional. 24 The windows used in (1.503) are the same as those introduced in note 20: the only difference is that they are
87
The choice of the window length is based on the compromise between spectral resolution and the variance of the estimate. An example has already been seen in the correlogram, where the condition L K =5 must be satised. Another example is the Welch periodogram. For a given observation of K samples, it is initially better to choose a small number of samples over which to perform the DFT, and therefore a large number of windows (subsequences) over which to average the estimate. The estimate is then repeated by increasing the number of samples per window, thus decreasing the number of windows. In this way we get estimates with a higher resolution, but also characterized by an increasing variance. The procedure is terminated once it is found that the increase in variance is no longer compensated by an increase in the spectral resolution. The aforementioned method is called window closing. Example 1.11.1 Consider a realization of K D 10000 samples of the signal: y.kTc / D
16 1 X h.nTc /w..k Ah nD 16
where '1 ; '2 2 U[0; 2 /, w.nTc / is a white random process with zero mean and variance 2 w D 5, Tc D 0:2, A1 D 1=20, f 1 D 1:5, A2 D 1=40, f 2 D 1:75, and Ah D Moreover sin .1 h.kTc / D / kTc T " C 4 kTc kTc cos .1 C / kTc T T rect # 8T C Tc kTc 2 kTc 4 T T
16 X 16
h.kTc /
(1.507)
(1.508)
with T D 4Tc and D 0:32. Actually y is the sum of two sinusoidal signals and ltered white noise through h. Consequently, observing (1.264) and (1.388),
2 P y . f / D w Tc
A2 jH. f /j2 C 1 .. f 2 4 Ah
f 1 / C . f C f 1 // (1.509)
A2 C 2 .. f 4
f 2 / C . f C f 2 //
where H. f / is the Fourier transform of fh.kTc /g. The shape of the PSD in (1.509) is shown in Figures 1.42 to 1.44 as a solid line. A Dirac impulse is represented by an isosceles triangle having a base equal to twice the desired
88
Figure 1.42. Comparison between spectral estimates obtained with Welch periodogram method, using the Hamming or the rectangular window, and the analytical PSD given by (1.509).
frequency resolution Fq . Consequently, a Dirac impulse, for example, of area A2 =4 will 1 have a height equal to A2 =.4Fq /, thus maintaining the equivalence in statistical power 1 between different representations. We now compare several spectral estimates, obtained using the previously described methods; in particular we will emphasize the effect on the resolution of the type of window used and the number of samples for each window. We state beforehand the following result. Windowing a complex sinusoidal signal f 1 /, fe j2 f 1 kTc g with fw.k/g produces a signal having Fourier transform equal to W. f where W. f / is the Fourier transform of w. Therefore, in the frequency domain the spectral line of a sinusoidal signal becomes a signal with shape W. f / centered around f 1 . In general, from (1.497), the periodogram of a real sinusoidal signal with amplitude A1 and frequency f 1 is 2 A1 Tc PPER . f / D jW. f f 1 / C W. f C f 1 /j2 (1.510) DMw 2 Figure 1.42 shows, in addition to the analytical PSD (1.509), the estimate obtained by the Welch periodogram method using the Hamming or the rectangular windows. Parameters used in (1.496) and (1.499) are: D D 1000, Ns D 19 and 50% overlap between windows. We observe that the use of the Hamming window yields an improvement of the estimate due to less leakage. Likewise Figure 1.43 shows how the Hamming window also improves the estimate carried out with the correlogram; in particular, the estimates of Figure 1.43 were obtained using in (1.503) L D 500. Finally, Figure 1.44 shows how the resolution and
89
Figure 1.43. Comparison between spectral estimates obtained with the correlogram using the Hamming or the rectangular window, and the analytical PSD given by (1.509).
Figure 1.44. Comparison of spectral estimates obtained with the Welch periodogram method, using the Hamming window, by varying parameters D ed Ns .
90
the variance of the estimate obtained by the Welch periodogram vary with the parameters D and Ns , using the Hamming window. Note that by increasing D, and hence decreasing Ns , both resolution and variance of the estimate increase.
1.12
ARMA(p,q) model
Let us consider the realization of a random process x according to the auto-regressive moving average model illustrated in Figure 1.45. In other words, the process x, also called observed sequence, is the output of an IIR lter having as input white noise with variance 2 w , and is given by the recursive equation25 x.k/ D
p X nD1
an x.k
n/ C
q X nD0
bn w.k
n/
(1.511)
Rewriting (1.511) in terms of the inputoutput relation of the linear system, from (1.129) we nd in general x.k/ D
C1 X nD0
h ARMA .n/w.k
n/
(1.512)
w(k)
Tc
w(k-1)
Tc
Tc
w(k-q)
b0
b1
bq
x(k)
ap x(k-p) Tc x(k-2)
a2 Tc x(k-1)
a1 Tc
25 In a simulation of the process, the rst samples x.k/ generated by (1.511) should be ignored because they
91
which indicates that the lter used to realize the ARMA model is causal. From (1.129) one nds that the lter transfer function is given by 8 q X > > B.z/ D bn z > > <
nD0 p > X > > A.z/ D > an z : nD0 n
HARMA .z/ D
B.z/ A.z/
where
(1.513)
n
assuming a0 D 1
Using (1.264), the power spectral density of the process x is given by: ( 2 B. f / D B.e j2 f Tc / 2 B. f / Px . f / D Tc w where A. f / A. f / D A.e j2 f Tc /
(1.514)
MA(q) model
If we particularize the ARMA model, assuming ai D 0 i D 1; 2; : : : ; p (1.515)
or A.z/ D 1, we get the moving average model of order q. The equations of the ARMA model therefore are reduced to HMA .z/ D B.z/ and
2 Px . f / D Tc w jB. f /j2
(1.516)
(1.517)
If we represent the function Px . f / of a process obtained by the MA model, we see that its behavior is generally characterized by wide peaks and narrow valleys, as illustrated in Figure 1.46.
AR(N) model
The auto-regressive model of order N is shown in Figure 1.47. The output process is described in this case by the recursive equation x.k/ D
N X nD1 2 where w is white noise with variance w . The transfer function is given by
an x.k
n/ C w.k/
(1.518)
HAR .z/ D
1 A.z/
(1.519)
92
w(k)
+ -
x(k)
aN
a2 Tc
a1 Tc Tc
with A.z/ D 1 C
N X nD1
an z
(1.520)
We observe that (1.519) describes a lter having N poles. Therefore HAR .z/ can be expressed as HAR .z/ D .1
1 /.1
p1 z
1 p2 z 1 / .1
pN z
1/
(1.521)
For a causal lter, the stability condition is jpi j < 1, i D 1; 2; : : : ; N , i.e. all poles must be inside the unit circle of the z plane.
93
In the case of the AR model, from Table 1.6 the z-transform of the ACS of x is given by Px .z/ D Pw .z/ 1 A.z/A D
2 w 1 A.z/A z
1 z
(1.522)
Hence the function Px .z/ has poles of the type jpi je j'i Letting A. f / D A.e j2 f Tc / one obtains the power spectral density of x, given by Px . f / D
2 Tc w
and
1 j'i e jpi j
(1.523)
(1.524)
jA. f /j2
(1.525)
Typically, the function Px . f / of an AR process will have narrow peaks and wide valleys (see Figure 1.48), reciprocal to the MA model.
94
(1.526)
For a given Px .z/, it is clear that the N zeros of A.z/ in (1.522) can be chosen in 2 N different ways. The selection of the zeros of A.z/ is called spectral factorization. Two examples are illustrated in Figure 1.49. As stated by the spectral factorization theorem (see page 53) there exists a unique spectral factorization that yields a minimum-phase A.z/, which is obtained by associating with A.z/ only the poles of Px .z/ that lie inside the unit circle of the z-plane.
Whitening lter
We observe an important property illustrated in Figure 1.50. Suppose x is modeled as an AR process of order N and has PSD given by (1.522). If x is input to a lter having transfer function A.z/, the output of this latter lter would be white noise. In this case the lter A.z/ is called whitening lter (WF). If A.z/ is minimum phase, the white process w is also called innovation of the process x, in the sense that the new information associated with the sample x.k/ is carried only by w.k/.
where s and x are uncorrelated processes. The process s, called predictable process, is described by the recursive equation s.k/ D
C1 X nD1
n s.k
n/
(1.528)
h.n/w.k
n/
(1.529)
95
Figure 1.49. Two examples of possible choices of the zeros () of A.z/, among the poles of Px .z/.
96
Theorem 1.4 (Kolmogorov theorem) Any ARMA or MA process can be represented by an AR process of innite order. Therefore any one of the three descriptions (ARMA, MA, or AR) can be adopted to approximate the spectrum of a process, provided that the order is sufciently high.
1.12.1
Autocorrelation of AR processes
It is interesting to evaluate the autocorrelation function of a process x obtained by the AR model. Multiplying both members of (1.518) by x .k n/, we nd x.k/x .k n/ D
N X mD1
am x.k
m/x .k
n/ C w.k/x .k
n/
(1.530)
Taking expectations, and observing that w.k/ is uncorrelated with all past values of x, for n 0 one gets E[x.k/x .k n/] D
N X mD1
am E[x.k
m/ x .k
2 n/] C w n
(1.531)
From (1.531), it follows rx .n/ D In particular we have 8 > > > > > > > > < > > > > > > > > :
N X mD1
am rx .n
2 m/ C w n
(1.532)
N X mD1 N X mD1
am rx .n
m/
rx .n/ D
2 am rx . m/ C w
r . x
n/
97
We observe that, for n > 0, rx .n/ satises an equation analogous to the (1.518), with the exception of the component w.k/. This implies that, if f pi g are zeros of A.z/, r x can be written as N X rx .n/ D ci pin n>0 (1.534)
i D1
(1.535)
Simplifying notation rx .n/ with r.n/, and observing (1.533), for n D 1; 2; : : : ; N , one gets a set of equations that in matrix notation are expressed as 3 2 3 2 a1 3 2 r.1/ r.0/ r. 1/ r. N C 1/ 6 6 r.2/ 7 7 6 r.1/ r.0/ r. N C 2/ 7 6 a2 7 7 6 76 6 (1.536) 76 : 7 D 6 : 7 6 : : :: 6 : 7 7 : : 54 : 5 4 : : : 4 : 5 : r.N 1/ r.N 2/ r.0/ r.N / aN that is Ra D r (1.537) with obvious denition of the vectors. In the hypothesis that the matrix R has an inverse, the solution for the coefcients fai g is given by a D R 1r (1.538) Equations (1.537) and (1.538), called YuleWalker equations, allow us to obtain the coef2 cients of an AR model for a process having autocorrelation function rx . The variance w of white noise at the input can be obtained from (1.533) for n D 0, which yields N X 2 am rx . m/ w D rx .0/ C (1.539) mD1 D rx .0/ C r H a Observation 1.8 From (1.537) one nds that a does not depend on rx .0/, but only on the correlation coefcients rx .n/ x .n/ D n D 1; : : : ; N (1.540) rx .0/ Exploiting the fact that R is Toeplitz and Hermitian, the set of equations (1.538) and (1.539) can be numerically solved by the LevinsonDurbin or by DelsarteGenin algorithms, with a computational complexity proportional to N 2 (see Sections 2.2.1 and 2.2.2). We note that the knowledge of rx .0/; rx .1/; : : : ; r x .N / univocally determines the ACS of an AR.N / process; for n > N , from (1.533), we get N X am rx .n m/ (1.541) rx .n/ D
mD1
98
1.12.2
Assuming an AR.N / model for a process x, (1.538) yields the coefcient vector a, which implies an estimate of the ACS up to lag N is available. From (1.525), we dene as spectral estimate PAR . f / D
2 Tc w
jA. f /j2
(1.542)
Usually the estimate (1.542) allows a better resolution than estimates obtained by other methods, such as PBT . f /, because it does not show the effects of ACS truncation. In fact the AR model yields PAR . f / D Tc
C1 X nD 1
rx .n/ e O
j2 f nTc
(1.543)
where rx .n/ is estimated for jnj N with one of the two methods of Section 1.11.2, while O for jnj > N the recursive equation (1.541) is used. The AR model accurately estimates processes with a spectrum similar to that given in Figure 1.48. For example, a spectral estimate for the process of Example 1.11.1 on page 87 obtained by an AR(12) model is depicted in Figure 1.51. The correlation coefcients were obtained by a biased estimate on 10000 samples. Note that the continuous part of the spectrum is estimated only approximately; on the other hand, the choice of a larger order N would result in an estimate with larger variance.
Figure 1.51. Comparison between the spectral estimate obtained by an AR(12) process model and the analytical PSD given by (1.509).
99
Also note that the presence of spectral lines in the original process leads to zeros of the polynomial A.z/ near the unit circle (see page 101). In practice, the correlation estimation method and a choice of a large N may result in an ill-conditioned matrix R. In this case the solution may have poles outside the unit circle, and hence the system would be unstable.
a1 rx .n
1/
n>0
D rx .0/ C a1 rx . 1/
2 w
(1.544)
ja1 j2
. a1 /jnj
(1.545)
j1 C a1 e
j2 f Tc j2
(1.546)
The behavior of the spectral density of an AR(1) process is illustrated in Figure 1.52.
100
AR(2). Let p1;2 D %e j'0 , where '0 D 2 f 0 Tc , be the two complex roots of A.z/ D 1 C a1 z 1 C a2 z 2 . We consider a real process: ( a1 D 2% cos.2 f 0 Tc / (1.547) a2 D % 2 Letting # D tan we nd s 2 2 1 C %2 1 %2 1C tan 1 .2 f 0 Tc / 2 2 % 1C% 2 1 %jnj cos.2 f 0 jnjTc rAR.2/ .n/ D w 1 %2 cos2 .4 f 0 Tc / C %4 The spectral density is thus given by PAR.2/ . f / D 1
2 Tc w 2 f 0 /Tc 1 j2. f C f 0 /Tc 2 1
1 %2 tan 1 C %2
.2 f 0 Tc /
(1.548)
#/ (1.549)
%e
j2. f
%e
(1.550)
We observe that, as % ! 1, PAR.2/ . f / has a peak that becomes more pronounced, as illustrated in Figure 1.53, and rx .k/ tends to exhibit a sinusoidal behavior.
101
Solutions of the YuleWalker equations are 8 rx .1/rx .0/ rx .1/rx .2/ >a D > 1 > > r2 .0/ r2 .1/ > x x > < 2 .1/ rx .0/rx .2/ rx > a2 D > > r2 .0/ r2 .1/ > x x > > 2 : w D rx .0/ C a1 rx .1/ C a2 rx .2/
(1.551)
Solving the previous set of equations with respect to rx .0/, rx .1/ and rx .2/, one obtains 8 2 w > > rx .0/ D 1 C a2 > > > 1 a2 .1 C a2 /2 a2 > 1 > > < a1 rx .1/ D rx .0/ (1.552) 1 C a2 > > ! > > > > a2 > r .2/ D 1 > x a2 C rx .0/ : 1 C a2 In general, for n > 0, we have " 2 p1 . p2 1/ rx .n/ D rx .0/ pn . p2 p1 /. p1 p2 C 1/ 1 # (1.553)
. p2
2 p2 . p1 1/ pn p1 /. p1 p2 C 1/ 2
with ' 2 U[0; 2 /. We observe that the process described by (1.554) satises the following difference equation for k 0: x.k/ D 2 cos.2 f 0 Tc /x.k 1/ x.k 2/ C k A cos ' k
1 A cos.2 f 0 Tc
'/ (1.555)
with x. 2/ D x. 1/ D 0. We note that the Kronecker impulses determine only the amplitude and phase of x. In the z-domain, we get the homogeneous equation A.z/ D 1 The zeros of A.z/ are p1;2 D e j2 f 0 Tc (1.557) 2 cos.2 f 0 Tc /z
1
Cz
(1.556)
It is important to verify that these zeros belong to the unit circle of the z plane. Consequently the representation of a sinusoidal process via the AR model is not possible, as the stability
102
condition, j pi j < 1, is not satised. Moreover the input (1.555) is not white noise. In any case, we can try to nd an approximation. In the hypothesis of uniform ', from Example 1.9.5, rx .n/ D A2 cos.2 f 0 nTc / 2 (1.558)
This autocorrelation function can be approximated by the autocorrelation of an AR(2) 2 process for % ! 1 and w ! 0. Using the formula (1.549), for % ' 1 we nd 2 6 rAR.2/ .n/ ' 6 4 2
2 w
2 1 %2
%2 cos2 .4 f
2 %!1;w !0
lim
(1.560)
Observation 1.9 We can observe the following facts about the order of an AR model approximating a sinusoidal process. From (1.390) one nds that an AR process of order N is required to model N complex sinusoids; on the other hand, from (1.388), one sees that an AR process of order 2N is required to model N real sinusoids. An ARMA process of order .2N ; 2N / is required to model N real sinusoids plus white 2 2 2 noise having variance b . Observing (1.513), it results w ! b and B.z/ ! A.z/. A better estimate is obtained by separating the continuous part from the spectral lines, for example by the scheme illustrated in Figure 3.38. The two components are then estimated by different methods.
1.13
Many of the topics surveyed in this chapter are treated in general in several texts on digital communications, in particular [4, 5, 6]. In-depth studies on deterministic systems and signals are found in [3, 7, 8, 9, 10, 11]. For a statistical analysis of random processes we refer to [1, 12, 13]. Finally, the subject of spectral estimation is discussed in detail in [2, 14, 15, 16].
1. Bibliography
103
Bibliography
[1] A. Papoulis, Probability, random variables and stochastic processes. New York: McGraw-Hill, 3rd ed., 1991. [2] M. B. Priestley, Spectral analysis and time series. New York, NY: Academic Press, 1981. [3] A. Papoulis, Signal analysis. New York: McGraw-Hill, 1984. [4] S. Benedetto and E. Biglieri, Principles of digital transmission with wireless applications. New York: Kluwer Academic Publishers, 1999. [5] D. G. Messerschmitt and E. A. Lee, Digital communication. Boston, MA: Kluwer Academic Publishers, 2nd ed., 1994. [6] J. G. Proakis, Digital communications. New York: McGraw-Hill, 3rd ed., 1995. [7] G. Cariolaro, La teoria unicata dei segnali. Turin: UTET, 1996. [8] A. Papoulis, The Fourier integral and its applications. New York: McGraw-Hill, 1962. [9] A. V. Oppenheim and R. W. Schafer, Discrete-time signal processing. Englewood Cliffs, NJ: Prentice-Hall, 1989. [10] P. P. Vaidyanathan, Multirate systems and lter banks. Englewood Cliffs, NJ: PrenticeHall, 1993. [11] R. E. Crochiere and L. R. Rabiner, Multirate digital signal processing. Englewood Cliffs, NJ: Prentice-Hall, 1983. [12] J. W. B. Davenport and W. L. Root, An introduction to the theory of random signals and noise. New York: IEEE Press, 1987. [13] A. N. Shiryayev, Probability. New York: SpringerVerlang, 1984. [14] S. M. Kay, Modern spectral estimation-theory and applications. Englewood Cliffs, NJ: Prentice-Hall, 1988. [15] L. S. Marple Jr., Digital spectral analysis with applications. Englewood Cliffs, NJ: Prentice-Hall, 1987. [16] P. Stoica and R. Moses, Introduction to spectral analysis. Englewood Cliffs, NJ: Prentice-Hall, 1997. [17] L. Erup, F. M. Gardner, and R. A. Harris, Interpolation in digital modemsPart II: implementation and performance, IEEE Trans. on Communications, vol. 41, pp. 998 1008, June 1993. [18] H. Meyr, M. Moeneclaey, and S. A. Fechtel, Digital communication receivers. New York: John Wiley & Sons, 1998.
104
Appendix 1.A
Multirate systems
1.A.1
Fundamentals
We consider the discrete-time linear transformation of Figure 1.54, with impulse response h.t/, t 2 <; the sampling period of the input signal is Tc , whereas that of the output signal is Tc0 . The inputoutput relation is given by the equation y.kTc0 / D
C1 X nD 1
h.kTc0
nTc /x.nTc /
(1.561)
We will use the following simplied notation: xn D x.nTc / yk D y.kTc0 / h.kTc0 If we assume that h has a nite support, say, between t1 and t2 , that is kTc0 or equivalently for n> then, letting n1 D kTc0 t2 Tc n< kTc0 t1 Tc (1.566) (1.567) (1.565) nTc < t2 kTc0 nTc > t1 (1.562) (1.563) nTc / 6D 0 for (1.564)
kTc0 t2 Tc
h.kTc0
n 1 Tc / C C xn 2 h.kTc0
n 2 Tc /
(1.568)
xn T c h
yk T c
105
One observes from (1.561) that: the values of h that contribute to yk are equally spaced by Tc ; the limits of the summation (1.568) are a complicated function of Tc , Tc0 , t1 , and t2 . Introducing the change of variable iD and setting 1k D kTc0 Tc t1 I1 D Tc t2 I2 D Tc kTc0 Tc t1 1k D Tc t2 1k D Tc (1.570) kTc0 C Tc 0 kTc kTc0 C Tc Tc kTc0 Tc (1.571) (1.572) kTc0 Tc n (1.569)
(1.568) becomes yk D
I2 X i DI1
(1.573)
From the denition (1.570) it is clear that 1k represents the truncation error of kTc0 =Tc and that 0 1k < 1. In the special case Tc0 M D Tc L with M and L integers, we get 1k D k M M k L L 1 M D kM k L L L D 1 .k M/mod L L (1.574)
(1.575)
We observe that 1k can assume L values f0; 1=L ; 2=L ; : : : ; .L 1/=Lg for any value of k. Hence there are only L univocally determined sets of values of h that are used in the computation of fyk g; in particular, if L D 1 only one set of coefcients exists, while
106
if M D 1 the sets are L. Summarizing, the output of a lter with impulse response h and with different input and output time domains can be expressed as yk D where gk;i D h..i C 1k /Tc / (1.577)
C1 X iD 1
gk;i xj k M k
L
(1.576)
We note that the system is linear and periodically time-varying. For Tc0 D Tc , that is for L D M D 1, we get 1k D 0, and the inputoutput relation is the usual convolution yk D
C1 X iD 1
g0;i x k
(1.578)
1.A.2
Decimation
Figure 1.55 represents a decimator or downsampler, with the output sequence related to the input sequence fxn g by yk D x k M (1.579)
where M, the decimation factor, is an integer number. We now obtain an expression for the z-transform of the output Y .z/ in terms of X .z/. We will show that Y .z/ D
2
X 1 1 M 1 m X .z M W M / M mD0
(1.580)
where W M D e j M is dened in (1.92). Equivalently, in terms of the radian frequency 0 normalized by the sampling frequency, !0 D 2 f =Fc , (1.580) can be written as Y .e j! / D
0
X !0 1 M 1 X ej M mD0
yk M
2 m M
(1.581)
xn Tc Fc = 1 Tc
T =MTc c Fc Fc = M
107
Figure 1.56. Decimation by a factor M D 3: a) in the time domain, and b) in the normalized radian frequency domain.
create M 1 replicas of the expanded version, and frequency-shift them uniformly with increments of 2 for each replica; sum all the replicas and divide the result by M. We observe that, after summation, the result is periodic in !0 with period 2 , as we would expect from a discrete Fourier transform. It is also useful to give the expression of the output sequence in the frequency domain; we get X m 1 M 1 (1.582) X f Y. f / D M mD0 M Tc where X . f / D X .e j2 f Tc / Y. f / D Y .e j2 f M Tc / (1.583) (1.584)
The relation (1.582) for the signal of Figure 1.56 is represented in Figure 1.57. Note that the only difference with respect to the previous representation is that all frequency responses are now functions of the frequency f . Proof of (1.580). The z-transform of fyk g can be written as Y .z/ D
C1 X kD 1
yk z
C1 X kD 1
x Mk z
(1.585)
108
k D 0; M; 2M; : : : otherwise
(1.586)
Y .z/ D
C1 X k0D 1
0 xk 0 M z
k0
C1 X kD 1
0 xk z
k=M
D X 0 .z 1=M /
(1.587)
This relation is valid, because x 0 is non-zero only at multiples of M. It only remains to express X 0 .z/ in terms of X .z/; to do this, we note that (1.586) can be expressed as
0 x k D ck x k
(1.588)
Note that the (1.589) can be written as ck D Hence we obtain X 0 .z/ D X C1 1 M 1 X x k W M km z M mD0 kD 1
k
X 1 M 1 W km M mD0 M X C1 1 M 1 X m x k zW M M mD0 kD 1
(1.590)
(1.591)
m The inner summation yields X .zW M /: hence, observing (1.587) we get (1.580).
109
1.A.3
Interpolation
Figure 1.58 represents an interpolator or upsampler, with the input sequence fxn g related to the output sequence by 8 > <x k k D 0; L ; 2L ; : : : L (1.592) yk D > :0 otherwise where L, the interpolation factor, is an integer number. We will show that the inputoutput relation in terms of the z-transforms Y .z/ and X .z/ is given by Y .z/ D X .z L / (1.593)
Equivalently, in terms of radian frequency normalized by the sampling frequency, !0 D 0 2 f =Fc , then (1.593) can be expressed as Y .e j! / D X .e j! L /
0 0
(1.594)
The graphical interpretation of (1.594) is illustrated in Figure 1.59: Y .e j! / is the compressed version by a factor L of X .e j! /; moreover, there are L 1 replicas of the compressed spectrum, called images. The creation of images implies that a lowpass signal does not remain lowpass after interpolation. It is also useful to give the expression of the output sequence in the frequency domain; we get Y. f / D X . f / where X . f / D X .e j2 f Tc / Tc Y. f / D Y e j2 f L (1.596) (1.597) (1.595)
The (1.595) for the signal of Figure 1.59 is illustrated in Figure 1.60. We note that the only 0 effect of the interpolation is that the signal X must be regarded as periodic with period Fc rather than Fc .
xn Tc 1 Fc = Tc L T = c yk Tc L
F =LF c c
110
Figure 1.59. Interpolation by a factor L D 3: (a) in the time domain, (b) in the normalized radian frequency domain.
yk z
C1 X nD 1
yn L z
nL
C1 X nD 1
xn z
nL
D X .z L /
(1.598)
1.A.4
Decimator lter
In most applications, a downsampler is preceded by a lowpass digital lter, to form a decimator lter as illustrated in Figure 1.61. The lter ensures that the signal vn is bandlimited, to avoid aliasing in the downsampling process.
111
Let h n D h.nTc /. Then we have yk D vk M and vn D The output can be expressed as yk D Using denition (1.577) we get gk;i D h i 8k; i (1.602)
C1 X iD 1 C1 X iD 1
(1.599)
h i xn
(1.600)
h i xk M
C1 X nD 1
hk M
n xn
(1.601)
Note that the overall system is not time invariant, unless the delay applied to the input is constrained to be a multiple of M. From V .z/ D X .z/H .z/ it follows that Y .z/ D X 1 M 1 m m H .z 1=M W M /X .z 1=M W M / M mD0 (1.603)
X 1 M 1 H .e M mD0
!0 2 m !0 2 m M /X .e j M /
(1.604)
1 0
j!j
M otherwise
(1.605)
we obtain Y .e j! / D
0
j!0 1 X eM M
j!0 j
(1.606)
In this case h is a lowpass lter that avoids aliasing caused by sampling; if x is bandlimited, the specications of h can be made less stringent. The decimator lter transformations are illustrated in Figure 1.62 for M D 4.
112
| X (f)|
| H (f)|
Fc /2
Fc f
| V (f)|
Fc /2
Fc f
| Y (f)|
Fc /2
Fc f
0 F /2 F c c
Fc /2
Fc
Figure 1.62. Frequency responses related to the transformations in a decimator lter for M D 4.
1.A.5
Interpolator lter
An interpolator lter is given by the cascade of an upsampler and a digital lter, as illustrated in Figure 1.63; the task of the digital lter is to suppress images created by upsampling [17]. Let h n D h.nTc0 /. Then we have the following inputoutput relations: yk D
C1 X jD 1
hk
jwj
(1.607)
wk D Therefore
8 < :
x 0
k L
k D 0; L ; : : : otherwise
(1.608)
yk D
C1 X rD 1
hk
r L xr
(1.609)
113
Let i D bk=Lc
gk;i xj k k
L
(1.610)
We note that gk;i is periodic in k of period L. In the z-transform domain we nd W .z/ D X .z L / Y .z/ D H .z/W .z/ D H .z/X .z L / or, equivalently, Y .e j! / D H .e j! /X .e j! L /
0 0 0
(1.611) (1.612)
(1.613)
where !0 D 2 f T =L D !=L. The interpolator lter transformations in the time and frequency domains are illustrated in Figure 1.64 for L D 3. If ( 1 j!0 j j!0 L (1.614) H .e / D 0 elsewhere we nd ( Y .e
j!0
/D
X .e j! / 0
j!0 j
L elsewhere
(1.615)
The relation between the input and output signal power for an interpolator lter is expressed by (1.419).
1.A.6
Rate conversion
Decimator and interpolator lters can be employed to vary the sampling frequency of a signal by an integer factor; in some applications, however, it is necessary to change the sampling frequency by a rational factor L=M. A possible procedure consists of rst converting a discrete-time signal into a continuous-time signal by a digital-to-analog converter (DAC), then re-sampling it at the new frequency. It is, however, easier and more convenient to change the sampling frequency by discrete-time transformations, for example, using the structure of Figure 1.65.
114
Figure 1.64. Time and frequency responses related to the transformations in an interpolator lter for L D 3.
115
This system can be thought of as the cascade of an interpolator and decimator lter, as illustrated in Figure 1.66, where h D h 1 h 2 . We obtain that ( ; 1 j!0 j min 0 L M H .e j! / D (1.616) 0 elsewhere In the time domain the following relation holds: C1 X yk D (1.617) gk;i xj k M k
iD 1 L i
where gk;i D h..i L C .k M/mod L /Tc0 / is the time-varying impulse response. In the frequency domain we get X !00 2l 1 M 1 00 V .e j M / Y .e j! / D M lD0 As we obtain Y .e j! / D From (1.616) we have
00
(1.618)
(1.619)
(1.620)
(1.621)
or
(1.622)
Example 1.A.1 (M > L: M D 5, L D 4) Transformations for M D 5 and L D 4 are illustrated in Figure 1.67. Observing the fact 0 that W .e j! / is zero for M !0 2 L M , the desired result is obtained by a response 0 H .e j! / that has the stopband cut-off frequency within this interval. Example 1.A.2 (M < L: M D 4, L D 5) The inverse transformation of the above example is obtained by a transformation with M D 4 and L D 5, as depicted in Figure 1.68.
116
X(e j )
W(e j )
4.2 4Fc
= 2 f T f
H(e j )
/L Fc /2
2 4Fc
= L f
V(e j )
/M LFc 2M
2 LFc
Y(e j ")
/M LFc 2M
2 LFc
f M=5
LFc 2M
5.2 LFc
" = M f
1.A.7
Time interpolation
Referring to the interpolator lter h of Figure 1.63, one nds that if L is large the lter implementation may require non-negligible complexity; in fact, the number of coefcients required for an FIR lter implementation can be very large. Consequently, in the case of a very large interpolation factor L, after a rst interpolator lter with a moderate value of the interpolation factor, the samples fyk D y.kTc0 /g may be further time interpolated until the desired sampling accuracy is reached [17]. As shown in Figure 1.69, let fyk g be the sequence that we need to interpolate to produce the signal z.t/, t 2 <; we describe below two time interpolation methods, linear and quadratic.
Linear interpolation
Given two samples yk 1 and yk , the signal z.t/, limited to interval [.k obtained by the linear interpolation z.t/ D y k
1
yk
1/
117
X(e j )
W(e j )
Fc /2
2 Fc
5.2 5Fc
= 2 f Tc f L=5
H(e j )
/5 Fc /2
2 5Fc
= L f
V(e j )
/5 Fc /2
2 LFc
Y(e j ")
/5 Fc /2
2 LFc
0 M L MFc Fc 2L 2
2 MFc L
. 42 LFc
" = M f
118
For an interpolation factor P of yk , we need to consider the sampling instants nTc00 D n and the values of z n D z.nTc00 / are given by .k 1/P (1.625) .yk yk 1 / P where n D .k 1/P; .k 1/P C 1; : : : ; k P 1. The case k D 1 is of particular interest: n n D 0; 1; : : : ; P 1 (1.626) z n D y0 C .y1 y0 / P In fact, regarding y0 and y1 as the two most recent input samples, their linear interpolation originates the sequence of P values given by (1.626). z n D yk
1
Tc0 P
(1.624)
Quadratic interpolation
In many applications linear interpolation does not always yield satisfactory results. Therefore, instead of connecting two points with a straight line, one resorts to a polynomial of degree Q 1 passing through Q points that are determined by the samples of the input sequence. For this purpose the Lagrange interpolation is widely used. As an example we report here the case of quadratic interpolation. In this case we consider a polynomial of degree 2 that passes through 3 points that are determined by the input samples. Let yk 1 , yk and ykC1 be the samples to interpolate by a factor P in the interval [.k 1/Tc0 ; .k C 1/Tc0 ]. The quadratic interpolation yields the values n0 n0 n0 n0 n0 n0 1C yk C 1 yk 1 C 1 C 1 ykC1 (1.627) zn D 2P P P P 2P P with n 0 D 0; 1; : : : ; P 1 and n D .k 1/ P C n 0 .
1.A.8
We recall some important properties of decimator and interpolator lters, known as noble identities; they will be used extensively in the next section on polyphase decomposition. Let G.z/ be a rational transfer function, i.e., a function expressed as the ratio of two polynomials in z or in z 1 ; it is possible to exchange the order of downsampling and ltering, or the order of upsampling and ltering as illustrated in Figure 1.70; in other words, the system of Figure 1.70a is equivalent to that of Figure 1.70b, and the system of Figure 1.70c is equivalent to that of Figure 1.70d. The proof of the noble identities is simple. For the rst identity, it is sufcient to note that W M m M D 1, hence Y2 .z/ D X 1 M 1 X .z 1=M W M m /G..z 1=M W M m / M / M mD0 (1.628) X 1 M 1 X .z 1=M W M m /G.z/ D Y 1 .z/ D M mD0
119
xn
G(z)
(a)
y1,k
xn
G(zM)
(b)
y2,k
xn
G(z)
(c)
y3,k
xn
x4,k
G(zL )
y4,k
(d)
For the second identity it is sufcient to observe that Y4 .z/ D G.z L /X 4 .z/ D G.z L /X .z L / D Y3 .z/ (1.629)
1.A.9
The polyphase representation allows considerable simplications in the analysis of transformations via interpolator and decimator lters, as well as the efcient implementation of such lters. To explain the basic concept, let us consider a lter having transfer function P1 H .z/ D nD0 h n z n . Separating the coefcients with even and odd time indices, we get H .z/ D Dening E .0/ .z/ D we can write H .z/ as H .z/ D E .0/ .z 2 / C z
1 1 X mD0 1 X mD0
h 2m z
2m
Cz
1 X mD0
h 2mC1 z
2m
(1.630)
h 2m z
E .1/ .z/ D
1 X mD0
h 2mC1 z
(1.631)
E .1/ .z 2 /
(1.632)
To expand this idea, let M be an integer; we can always decompose H .z/ as H .z/ D
1 X mD0
hm M z
1 1 X mD0
mM
Cz Letting
h m MC1 z
mM
C C z
.M 1/
1 X mD0
(1.633) h m MCM
1z mM
.`/ em D h m MC`
0`M
(1.634)
120
we can express compactly the previous equation as H .z/ D where E .`/ .z/ D
1 X i D0 M 1 X `D0
E .`/ .z M /
(1.635)
ei.`/ z
(1.636)
The expression (1.635) is called the type 1 polyphase representation (with respect to M), and E .`/ .z/, where ` D 0; 1; : : : ; M 1, the polyphase components of H .z/. The polyphase representation of an impulse response fh n g with 7 coefcients is illustrated in Figure 1.71 for M D 3. A variation of (1.635), called type 2 polyphase representation, is given by H .z/ D
M 1 X `D0
.M 1 `/
R .`/ .z M /
(1.637)
1 `/ .z/.
where the components R .`/ .z/ are permutations of E .`/ .z/, that is R .`/ .z/ D E .M
Efcient implementations
The polyphase representation is the key to obtaining efcient implementation of decimator and interpolator lters. In the following, we will rst consider the efcient implementations for M D 2 and L D 2, then we will extend the results to the general case.
121
Figure 1.72. Implementation of a decimator lter using the type 1 polyphase representation for M D 2.
Figure 1.73. Optimized implementation of a decimator lter using the type 1 polyphase representation for M D 2.
Decimator lter. Referring to Figure 1.61, we consider a decimator lter with M D 2. By (1.635), we can represent H .z/ as illustrated in Figure 1.72; by the noble identities, the lter representation can be drawn as in Figure 1.73a. The structure can be also drawn as in Figure 1.73b, where input samples fxn g are alternately presented at the input to the two lters e .0/ and e .1/ ; this latter operation is generally called serialtoparallel (S/P) conversion. Note that the system output is now given by the sum of the outputs of two lters, each operating at half the input frequency and having half the number of coefcients as the original lter. To formalize the above ideas, let N be the number of coefcients of h, and N .0/ and .1/ be the number of coefcients of e .0/ and e .1/ , respectively, so that N D N .0/ C N .1/ . N In this implementation e .`/ requires N .`/ multiplications and N .`/ 1 additions; the total cost is still N multiplications and N 1 additions, but, as e .`/ operates at half the input rate, the computational complexity in terms of multiplications per second (MPS) is N Fc 2 while the number of additions per second (APS) is given by MPS D APS D .N (1.638)
1/Fc (1.639) 2 Therefore the complexity is about one half the complexity of the original lter. The efcient implementation for the general case is obtained as an extension of the case for M D 2 and is shown in Figure 1.74.
122
Figure 1.74. Implementation of a decimator lter using the type 1 polyphase representation.
Figure 1.75. Implementation of an interpolator lter using the type 1 polyphase representation for L D 2.
Interpolator lter. With reference to Figure 1.63, we consider an interpolator lter with L D 2. By (1.635), we can represent H .z/ as illustrated in Figure 1.75; by the noble identities, the lter representation can be drawn as in Figure 1.76a. The structure can be also drawn as in Figure 1.76b, where output samples are alternately taken from the output of the two lters e .0/ and e .1/ ; this latter operation is generally called paralleltoserial (P/S) conversion. Remarks on the computational complexity are analogous to those of the decimator lter case. In the general case, efcient implementations are easily obtainable as extensions of the case for L D 2 and are shown in Figure 1.77. The type 2 polyphase implementations of interpolator lters are depicted in Figure 1.78. Interpolator-decimator lter. As illustrated in Figure 1.79, at the receiver of a transmission 0 system it is often useful to interpolate the signal fr.nTQ /g from TQ to TQ to get the signal 0 /g. fx.q TQ Let rn D r.nTQ /
0 xq D x.q TQ /
(1.640)
123
P/S
xn
E (0) (z)
xn Tc z -1
E (1) (z)
2
(a)
yk
T = Tc c 2
Figure 1.76. Optimized implementation of an interpolator lter using the type 1 polyphase representation for L D 2.
Figure 1.77. Implementation of an interpolator lter using the type 1 polyphase representation.
0 The sequence fx.q TQ /g is then downsampled with timing phase t0 . Let yk be the output with sampling period Tc ,
(1.641)
(1.642)
t0 0 D `0 C L0 L TQ
(1.643)
where `0 2 f0; 1; : : : ; L 1g, and L0 is a non-negative integer number. For the general 0 case of an interpolator-decimator lter where t0 and the ratio Tc =TQ are not constrained, we refer to [18] (see also Chapter 14).
124
xn
R(0)(z)
xn L z-1
R(0)(z)
P/S k=L-1
R(1)(z)
L z-1
R(1)(z)
k=L-2 yk
R(L-1) (z)
(a)
yk L
R(L-1) (z)
k=0
(b)
Figure 1.78. Implementation of an interpolator lter using the type 2 polyphase representation.
(1.644)
fE .`/ .z/g
` D 0; 1; : : : ; L
(1.645)
0 The interpolator lter structure from TQ to TQ is illustrated in Figure 1.77. For the special case M D 1, that is for Tc D TQ , the implementation of the interpolator-decimator lter is given in Figure 1.80, where
yk D vkCL0
(1.646)
125
Figure 1.80. Polyphase implementation of an interpolator-decimator lter with timing phase t0 D .`0 C L0 L/T0 . Q
Figure 1.81. Implementation of an interpolator-decimator lter with timing phase t0 D 0 .`0 C L0 L/TQ .
126
In other words, fyk g coincides with the signal fvn g at the output of branch `0 of the polyphase structure. In practice we need to ignore the rst L0 samples of fvn g, as the relation between fvn g and fyk g must take into account a lead, z L0 , of L0 samples. With 0 reference to Figure 1.80, the output fxq g at instants that are multiples of TQ is given by the outputs of the various polyphase branches in sequence. In fact, let q D ` C n L, ` D 0; 1; : : : ; L 1, and n integer, we have
0 0 0 x`Cn L D x.n L TQ C `TQ / D x.nTQ C `TQ /
(1.647)
We now consider the general case M 6D 1. First, to downsample the signal interpolated 0 at TQ one can still use the polyphase structure of Figure 1.80. In any case, once t0 is chosen, the branch is identied (say `0 ) and its output must be downsampled by a factor M L. Notice that there is the timing lead L0 L in (1.643) to be considered. Given L0 , we determine a positive integer N0 so that L0 C N0 is a multiple of M, that is L0 C N0 D M0 M (1.648)
The structure of Figure 1.80, considering only branch `0 , is equivalent to that given in Figure 1.81a, in which we have introduced a lag of N0 samples on the sequence frn g and a further lead of N0 samples before the downsampler. In particular we have r 0p D r p N0 and x 0p D x p N0 (1.649)
As a result, the signal is not modied before the downsampler. Using now the representation of E .`0 / .z L / in M phases: E .`0 ;m/ .z L M / m D 0; 1; : : : ; M 1 (1.650)
127
Appendix 1.B
Let w D w I C j w Q be a complex Gaussian r.v. with zero mean and unit variance; note N N N N N N that w I D Re [w] and w Q D Im [w]. In polar notation, N w D A e j' N (1.651)
It can be shown that ' is a uniform r.v. in [0; 2 /, and A is a Rayleigh r.v. with probability distribution ( 2 1 e a a>0 P[A a] D (1.652) 0 a<0 Observing (1.652) and (1.651), if u 1 and u 2 are two uniform r.v.s in the interval [0; 1/, then p ln.1 u 1 / (1.653) AD and ' D 2 u 2 (1.654) In terms of real components, it results that w I D A cos ' N and w Q D A sin ' N (1.655)
are two statistically independent Gaussian r.v.s, each with zero mean and variance equal to 0.5. The r.v. w is also called circularly symmetric Gaussian r.v., as the real and imaginary N components, being statistically independent with equal variance, have a circularly symmetric Gaussian joint probability density function.
Chapter 2
The theory of the Wiener lter [1, 2] that will be presented in this chapter is fundamental to the comprehension of several important applications. The development of this theory assumes the knowledge of the correlation functions of the relevant processes. An approximation of the Wiener lter can be obtained by least squares methods, through realizations of the processes involved.
2.1
With reference to Figure 2.1, let x and d be two individually and jointly wide sense stationary random processes with zero mean; the problem is to determine the FIR lter so that, if the lter input is x.k/, the output y.k/ replicates as closely as possible d.k/. The Wiener theory provides the means to design the required lter. The FIR lter in Figure 2.1 is called transversal lter, as the output is formed by summing the products of the delayed input samples by suitable coefcients. If we indicate with fcn g, n D 0; 1; : : : ; N 1, the N coefcients of the lter, we have y.k/ D
N 1 X nD0
cn x.k
n/
(2.1)
If d.k/ is the desired sample at the lter output at instant k, we dene the estimation error as e.k/ D d.k/ y.k/ (2.2)
In the Wiener theory, to cause the lter output y.k/ to replicate as closely as possible d.k/, the coefcients of the lter are determined using the minimum mean-square error (MMSE) criterion. Therefore the cost function is dened as J D E[je.k/j2 ] and the coefcients of the optimum lter are those that minimize J :
fcn g;nD0;1;:::;N 1
(2.3)
min
(2.4)
130
The Wiener lter problem can be formulated as the problem of estimating d.k/ by a linear combination of x.k/; : : : ; x.k N C 1/. A brief introduction to estimation theory is given in Appendix 2.A; in the second half of the Appendix, of which reading should be deferred until the end of this section, the formulation of the Wiener theory is further extended to the case of vector signals.
Matrix formulation
The problem introduced in the previous section is now formulated using matrix notation. We dene:1 1. Coefcient vector cT D [c0 ; c1 ; : : : ; c N 2. Filter input vector at instant k xT .k/ D [x.k/; x.k The lter output at instant k is expressed as y.k/ D cT x.k/ D xT .k/ c and the estimation error as e.k/ D d.k/ cT x.k/ (2.8) (2.7) 1/; : : : ; x.k N C 1/] (2.6)
1]
(2.5)
The components of an N -dimensional vector are usually identied by an index varying either from 1 to N or from 0 to N 1.
131
We express now the cost function J as a function of the vector c. We will then seek that particular vector copt that minimizes J . Recalling the denition J D E[e.k/e .k/] D E[.d.k/ and computing the products, it follows J D E[d .k/d.k/] c H E[x .k/d.k/] C (2.11) xT .k/c/.d .k/ c H x .k//] (2.10)
Assuming that x and d are individually and jointly WSS, we introduce the following quantities: 1. Variance of the desired signal
2 d D E[d.k/d .k/]
(2.12)
2. Correlation between the desired output at instant k and the lter input vector at the same instant 2 3 E[d.k/x .k/] 6 7 6 E[d.k/x .k 1/] 7 6 7 7Dp rdx D E[d.k/x .k/] D 6 (2.13) 6 7 : : 6 7 : 4 5 E[d.k/x .k N C 1/] The components of p are given by [p]n D E[d.k/x .k Moreover, it holds p H D E[d .k/xT .k/] 3. N N correlation matrix of the lter input vector, as dened in (1.346), R D E[x .k/xT .k/] Then
2 J D d
n D 0; 1; : : : ; N
(2.14)
(2.15)
(2.16)
cH p
p H c C c H Rc
(2.17)
The cost function J , considered as a function of c, is a quadratic function. Then, if R is positive denite, J admits one and only one minimum value. Plots of J are shown in Figure 2.2, for the particular cases N D 1 and N D 2.
132
In general, because the vector p and the autocorrelation matrix R are complex, we also write p D p I C jp Q (2.20)
133
and R D R I C jR Q If now we take the derivative of the terms of (2.17) using (2.19), we nd rc I p H c D rc I p H .c I C jc Q / D p H rc Q p H c D rc Q p H .c I C jc Q / D jp H rc I c H p D rc I .cT I rc Q c H p D rc Q .cT I rc I .c H Rc/ D 2R I c I jcT /p D p Q jcT /p D Q 2R Q c Q jp (2.22) (2.23) (2.24) (2.25) (2.26) (2.27) (2.21)
rc Q .c H Rc/ D 2R I c Q C 2R Q c I From the above equations we obtain rc p H c D 0 rc c p D 2p rc c H Rc D 2Rc Substituting the above results into (2.17), it turns out rc J D 2p C 2Rc D 2.Rc p/
H
(2.31)
For the optimum coefcient vector copt the components of rc J are all equal to zero, hence we get Rcopt D p Then (2.32) is called the WienerHopf equation (W-H). Observation 2.1 The computation of the optimum coefcients copt requires the knowledge only of the input correlation matrix R and of the cross-correlation vector p between the desired output and the input vector. In scalar form, the WienerHopf equation is a system of N equations in N unknowns:
N 1 X i D0
(2.32)
copt;i rx .n
i/ D rdx .n/
n D 0; 1; : : : ; N
(2.33)
If R
(2.34)
134
Corollary 2.1 For c D copt , e.k/ and y.k/ are orthogonal: E[e.k/y .k/] D 0 In fact, using the orthogonality principle, E[e.k/y .k/] D E[e.k/c H x .k/] D c H E[e.k/x .k/] D cH 0 D0 For an optimum lter, Figure 2.3 depicts the relation between the three signals d.k/, e.k/, and y.k/. (2.39) for c D copt (2.38)
d(k) e(k)
y(k)
Figure 2.3. Orthogonality of signals for an optimum lter.
Note that orthogonality holds only if e and x are considered at the same instant. In other words, the notion of orthogonality between random variables is used.
135
p H copt
(2.40)
Another useful expression of Jmin is obtained from (2.2): d.k/ D e.k/ C y.k/ As e.k/ and y.k/ are orthogonal for c D copt , then
2 2 d D Jmin C y
(2.41)
(2.42)
whereby it follows
2 Jmin D d 2 y
(2.43)
Using (2.40) we can nd an alternative expression to (2.17) for the cost function J : J D Jmin C .c copt / H R.c copt / (2.44)
Recalling that the autocorrelation matrix is positive semi-denite, it follows that the quantity .c copt / H R.c copt / is non-negative and in particular it vanishes for c D copt .
3 1 6 : 7 D 4 : 5 D U H .c : N
copt /
(2.46)
where i D uiH .c copt /. The vector may be interpreted as a translation and a rotation of the vector c. Then J assumes the form: J D Jmin C H D Jmin C
N X i D1 N X i D1
i ji j2
D Jmin C
i juiH .c
136
The result (2.47) expresses the excess mean-square error J Jmin as the sum of N components in the direction of each eigenvector of R. Note that each component is proportional to the corresponding eigenvalue. The above observation allows us to deduce that J increases more rapidly in the direction of the eigenvector corresponding to the maximum eigenvalue max . Likewise the increase is slower in the direction of the eigenvector corresponding to the minimum eigenvalue min . Let umax and umin denote the eigenvalues of R in correspondence of max and min , respectively; it follows that rc J is largest along umax . This is also observed in Figure 2.4, where sets (loci) of points c for which a constant value of J is obtained are graphically represented. In the 2-dimensional case they trace ellipses with axes that are parallel to the direction of the eigenvectors and ratio of axes that is related to the value of the eigenvalues.
copt;i rx .n
i/ D rdx .n/
8n
(2.48)
Taking the z-transform of both members yields Copt .z/Px .z/ D Pdx .z/ Then the transfer function of the optimum lter is given by Copt .z/ D Pdx .z/ Px .z/ (2.50) (2.49)
We note that while (2.34) is useful in evaluating the coefcients of the optimum FIR lter, equation (2.50) is employed to analyze the system in the general case of an IIR lter. Example 2.1.1 Let d.k/ D h x.k/, as shown in Figure 2.5. In this case, from Table 1.3, Pdx .z/ D Px .z/H .z/ (2.51)
137
d(k) +
c x(k)
y(k) e(k)
The optimum lter is given by Copt .z/ D H .z/ From (2.40) in scalar notation, applying Fourier transform properties, we get
2 Jmin D d N 1 X i D0
(2.52)
2 D d
1 2Tc 1 2Tc
1 2Tc 1 2Tc
Pdx . f /
Pdx . f / df Px . f /
2 d
1 2Tc 1 2Tc
(2.54)
2 d
Z Tc
1 2Tc 1 2Tc
Example 2.1.2 We want to lter the noise from a signal given by one complex sinusoid (tone) plus noise, i.e. x.k/ D Ae j .!0 kC'/ C w.k/ (2.55)
In (2.55) !0 D 2 f 0 Tc is the tone radian frequency normalized to the sampling period, in radians. We assume the desired signal is given by d.k/ D B e j[!0 .k
D/C']
(2.56)
where D is a known delay. We also assume that ' 2 U.0; 2 /, and w is white noise with 2 zero mean and variance w , uncorrelated with '. The autocorrelation function of x and the
138
(2.57) (2.58)
D/
For a Wiener lter with N coefcients, the autocorrelation matrix R and the vector p have the following structure: 2 6 6 RD6 6 4 2 6 6 pD6 6 4
2 A 2 C w
A2 e
j!0
: : : A2 e
j!0 .N 1/ 2/
3 7 7 7 7 5 (2.59)
2 A 2 C w : : :
: : : A2 e j!0 .N :: : :::
2/
A2 e j!0 .N
:::
2 A 2 C w
3 7 7 7 AB e 7 5
j!0 D
(2.60)
(2.61)
(2.62) (2.63)
p D ABe
j!0 D
E.!0 /
Observing that E H .!/E.!/ D N , the inverse of R is given by # " 1 A2 E.!0 /E H .!0 / R 1D 2 I 2 w w C N A 2 Hence, using (2.34): copt D ABe j!0 D B 3e j!0 D E.!0 / E.!0 / D 2 A 1 C N3 w C N A 2
(2.64)
(2.65)
2 where 3 D A2 =w is the signal-to-noise ratio. From (2.40) the minimum value of the cost function J is given by
Jmin D B 2
ABe j!0 D B2 D 2 1 C N3 w C N A 2
(2.66)
139
copt;i e
j!i
(2.67)
! D !0
j .! !0 /N j .! !0 /
(2.68) ! 6D !0
E.!0 /;
3. jCopt .e j!0 /j D
B . A
Figure 2.6. Magnitude of Copt .e j2 fTc / given by (2.68) for f0 Tc D 1=2, B D A, 3 D 30 dB, and N D 35.
140
Conversely, for 3 ! 0, i.e. when the power of the useful signal is negligible with respect to the power of the additive noise, it results in 1. Jmin D B 2 ; 2. copt D 0; 3. jCopt .e j!0 /j D 0. Indeed, as the signal-to-noise ratio vanishes the best choice is to set the output y to zero. The plot of jCopt .e j2 f Tc /j is given in Figure 2.6.
2.2
Linear prediction
The Wiener theory considered in the previous section has an important application to the solution of the following problem. Let x be a discrete-time WSS random process with zero mean; prediction consists in estimating a future value of the process starting from a set of known past values. In particular, let us dene the vector xT .k 1/ D [x.k 1/; x.k 2/; : : : ; x.k N /] (2.69)
The one-step forward predictor of order N , given xT .k 1/, attempts to estimate the value of x.k/. There exists also the problem of predicting x.k N /, given the values of x.k N C 1/; : : : ; x.k/. In this case, the system is called the one-step backward predictor of order N .
ci x.k
i/
(2.70)
Tc
x(k-1)
Tc
x(k-2)
Tc
c N-1
x(k-N)
c1
c2
cN
^ x(k| x (k-1) )
Figure 2.7. Linear predictor of order N.
141
This estimate will be subject to a forward prediction error given by f N .k/ D x.k/ D x.k/ x.k j x.k O
N X i D1
1// (2.71) i/
ci x.k
to determine the predictor coefcients, we can use the optimization results according to Wiener. We recall the following denitions. 1. Desired signal d.k/ D x.k/ 2. Filter input vector (dened by (2.69)) xT .k 3. Cost function J (given by (2.72)). Then it turns out:
2 2 d D E[x.k/x .k/] D x D rx .0/
(2.73)
1/
(2.74)
(2.75) (2.76)
E[x .k
1/x .k
T
p D E[d.k/x .k
1/] D E[x.k/x .k
(2.77)
Applying (2.32), the optimum coefcients satisfy the equation R N copt D r N Moreover, from (2.40) we get the minimum value of the cost function J , Jmin D J N D rx .0/
H r N copt
(2.78)
(2.79)
We can combine the latter two equations to get an augmented form of the W-H equation for the linear predictor: " #" # " # H 1 JN rx .0/ r N D (2.80) 0N copt rN RN where 0 N is the column vector of N zeros.
142
copt;i x.k
i/
(2.81)
1 copt;i D
i D0 i D 1; 2; : : : ; N
(2.82)
where a D copt . Substituting (2.82) in (2.81) and taking care to extend the equation also to i D 0, we obtain f N .k/ D
N X i D0 0 ai;N x.k
i/
(2.84)
as shown in Figure 2.8. 0T The coefcients a N D [a0 ; a0 ; : : : ; a0N ;N ] are directly obtained by substituting 0;N 1;N (2.83) in (2.80): JN 0 (2.85) R N C1 a N D 0N
x(k)
Tc
x(k-1)
Tc
x(k-2)
Tc
x(k-N)
a 0,N
a 1,N
a 2,N
a N-1,N
a N,N
f (k)
N
143
With a similar procedure, we can derive the lter that gives the backward linear prediction error, b N .k/ D x.k N/
N X i D1
gi x.k
i C 1/
(2.86)
(2.87)
where B is the backward operator that orders the elements of a vector backward, from the last to the rst (see page 27).
144
it results 8 > a0 D J1 r .0/ > > 0;1 x < 1r > > 0 >a D : 1;1 where r .0/ r .1/ x x 1r D D r2 .0/ x rx .1/ rx .0/ As a0 D 1, it turns out 0;1 8 > J D 1r > 1 > < rx .0/ > > 0 >a : N D 2. 8 > a0 D > 1;2 > < > > 0 >a D : 2;2 and J2 1 D rx .0/ 2j.1/j2 C 2 .1/.2/ C j.1/j2 .2/ 1 j.1/j2 j.2/j2 (2.93)
1;1
J1 rx .1/ 1r
(2.89)
jrx .1/j2
(2.90)
rx .1/ rx .0/
J1 D1 rx .0/
rx .1/rx .0/ r .1/rx .2/ x r2 .0/ jrx .1/j2 x rx .0/rx .2/ r2 .1/ x r2 .0/ jrx .1/j2 x
.1/.2/ j.1/j2
We note that in the above equations .n/ is the correlation coefcient of x, introduced in (1.540).
145
2.2.1
The LevinsonDurbin algorithm (LDA) yields the solution of matrix equations like (2.85), in which R N C1 is positive denite, Hermitian, and Toeplitz, with a computational complexity proportional to N 2 , instead of N 3 as happens with algorithms that make use of the inverse matrix. In the case of real signals, R N C1 is symmetric and the computational complexity of the DelsarteGenin algorithm (DGA), given in Section 2.2.2, is halved with respect to that of LDA. Here, we report a step-by-step description of the LDA: 1. Initialization. We set: J0 D rx .0/ 10 D rx .1/ 2. n-th iteration, n D 1; 2; : : : ; N . We calculate Cn D "
0 an D
(2.94) (2.95)
1n Jn
1 1
(2.96) # C Cn " 0 an
0 B
0 an 1 0
# (2.97)
1
Then (2.97) corresponds to the scalar equations: a0 D a0 k;n k;n with a0 0;n
1 0 D 1 and an;n 1 1
C Cn an
k;n 1
k D 0; 1; : : : ; n
(2.98)
D 0. Moreover,
B 0 1n D .rnC1 /T an
(2.99) (2.100)
Jn D Jn
1 .1
jCn j2 /
We now interpret the physical meaning of the parameters in the algorithm. Jn represents the statistical power of the forward prediction error at the n-th iteration: Jn D E[j f n .k/j2 ] It results in 0 Jn Jn with J0 D rx .0/ and J N D J0
N Y nD1 1
(2.101)
n1
(2.102)
(2.103)
.1
jCn j2 /
(2.104)
146
D E[ f n
1 .k/bn 1 .k
1/]
(2.105)
In other words, 1n can be interpreted as the cross-correlation between the forward linear prediction error and the backward linear prediction error delayed by one sample. Cn satises the following property:
0 Cn D an;n
(2.106)
Finally, by substitution, from (2.96), along with (2.101) and (2.105), we get Cn D and, noting that E[j f n
2 1 .k/j ]
E[ f n
1/]
1 .k
D E[jbn
2 1 .k/j ]
D E[jbn
jCn j 1
The coefcients fCn g are called reection coefcients or partial correlation coefcients (PARCOR).
Lattice lters
We have just described the LevinsonDurbin algorithm. Its analysis permits us to implement the prediction error lter via a modular structure. Dening xnC1 .k/ D [x.k/; : : : ; x.k we can write: " xnC1 .k/ D xn .k/ x.k n/ # D " x.k/ xn .k 1/ # (2.110) n/]T (2.109)
We recall the relation for forward and backward linear prediction error lters of order n: 8 0 0 < f n .k/ D a0 x.k/ C C an;n x.k n/ D anT xnC1 .k/ 0;n (2.111) : b .k/ D a0 x.k/ C C a0 x.k n/ D a0 B H x .k/ n nC1 n;n n 0;n From (2.97) we obtain " f n .k/ D D fn
0 an 1 0
#T xnC1 .k/ C Cn
"
0 an
0 B
#T xnC1 .k/
1
(2.112)
1 .k/ C Cn bn 1 .k
1/
(2.113)
147
f0 (k)
f1 (k)
f m-1 (k)
fm (k)
f N-1 (k)
fN (k) CN
C1 x(k)
Cm
C* 1
b0 (k)
* Cm
* CN
Tc
b1 (k)
bm-1 (k)
Tc
bm (k)
bN-1 (k)
Tc
bN (k)
Finally, taking into account the initial conditions, f 0 .k/ D b0 .k/ D x.k/ and a0 D 1 0;0 the block diagram of Figure 2.10 is obtained, in which the output is given by f N . We list the following fundamental properties. 1. The optimum coefcients Cn , n D 1; : : : ; N , are independent of the order of the lter; therefore one can change N without having to re-calculate all the coefcients. This property is useful if the lter length is unknown and must be estimated. 2. If the conditions jCn j 1, n D 1; : : : ; N , are veried, the lter is minimum phase. 3. The lattice lters are quite insensitive to coefcient quantization. Observation 2.2 From the above property 2 and (2.108), we nd that all predictor error lters are minimum phase. (2.114)
2.2.2
In the case of real signals, the DGA, also known as the split Levinson algorithm [3], further reduces the number of operations with respect to the LDA, at least for N 10.3 Here is the step-by-step description. 1. Initialization. We set v0 D 1 v1 D [1; 1]T 0 D rx .0/ 1 D rx .0/ C rx .1/
0 1
(2.115) (2.116)
Faster algorithms, with a complexity proportional to N .log N /2 , have been proposed by Kumar [4].
148
(2.117)
2
n D 2n
n n C
vn 1 0
0 vn
1
0 n 4 vn 2 5 0
(2.118) (2.119)
n D
n n 1 n n n
(2.121) 0 vn
n 1 1
0 an D vn
Jn D n Cn D 1
We note that (2.120) exploits the symmetry of the vector vn ; in particular it is [vn ]1 D [vn ]nC1 D 1.
2.3
The Wiener lter will prove to be a powerful analytical tool in various applications, one of which is indeed prediction. However, from a practical point of view, often only realizations of the processes fx.k/g and fd.k/g are available. Therefore to get the solution it is necessary to determine estimates of rx and rdx , and various alternatives emerge. Two possible methods are: 1) the autocorrelation method, in which from the estimate of rx we construct R as a Toeplitz correlation matrix, and 2) the covariance method, in which we estimate each element of R by (2.130). In this case the matrix is only Hermitian and the solution that we are going to illustrate is of the LS type [1, 2]. We reconsider the problem of Section 2.1, introducing a new cost function. Based on the observation of the sequences fx.k/g and of the error e.k/ D d.k/ y.k/ (2.126) k D 0; : : : ; K 1 and fd.k/g k D 0; : : : ; K 1 (2.125)
where y.k/ is given by (2.1), according to the least squares method the optimum lter coefcients yield the minimum of the sum of the squared errors:
fcn g;nD0;1;:::;N 1
min
(2.127)
149
where ED
K 1 X kDN 1
je.k/j2
(2.128)
Note that in the LS method a time average is substituted for the expectation (2.3), which gives the MSE.
Data windowing
In matrix notation, the output fy.k/g, k D N expressed as 3 2 2 x.N 1/ x.N 2/ y.N 1/ 6 y.N / 7 6 x.N / x.N 1/ 7 6 6 7D6 6 : : : : : : 5 4 4 : : : y.K 1/ | x.K 1/ x.K 1; : : : ; K ::: ::: :: : x.0/ x.1/ : : : 1, given by (2.1), can be 32 76 76 76 54 } c0 c1 : : : cN
1
3 7 7 7 5 (2.129)
data matrix T
2/ : : : x.K {z
N/
In (2.129) we note that the input data sequence actually used goes from x.0/ to x.K 1/. Other choices are possible for the input data window. The case examined is called the covariance method and the data matrix T, dened by (2.129), is LN where L D K N C1.
Matrix formulation
We dene 8.i; n/ D
K 1 X kDN 1
x .k
i/ x.k
n/
i; n D 0; 1; : : : ; N
(2.130)
#.n/ D
K 1 X kDN 1
d.k/ x .k
n/
n D 0; 1; : : : ; N
(2.131)
Using (1.478) for an unbiased estimate of the correlation, the following identities hold: 8.i; n/ D .K #.n/ D .K N C 1/rx .i O N C 1/rdx .n/ O n/ (2.132) (2.133)
in which the values of 8.i; n/ depend on both indices .i; n/ and not only upon their difference, especially if K is not very large. We give some denitions: 1. Energy of fd.k/g Ed D
K 1 X kDN 1
jd.k/j2
(2.134)
150
2. Cross-correlation vector between d and x T D [#.0/; #.1/; : : : ; #.N 3. Input autocorrelation matrix 2 8.0; 0/ 6 8.1; 0/ 6 D6 : : 4 : 8.N 1/] (2.135)
1; 0/ 8.N
8.0; 1/ 8.1; 1/ : : :
::: ::: :: :
1; 1/ : : : 8.N
8.0; N 8.1; N : : : 1; N
1/ 1/ 1/
3 7 7 7 5 (2.136)
Correlation matrix
is the time average of x .k/xT .k/, i.e. D
K 1 X kDN 1
x .k/xT .k/
(2.138)
Properties of 1. 2.
3. Eigenvalues of 4.
with T input data matrix dened by (2.129). We note that the matrix T is Toeplitz.
Then the vector of optimum coefcients based on the LS method, cls , satises the normal equation cls D (2.141)
151
If
(2.142)
In the solution of the LS problem, the equation (2.141) corresponds to the WienerHopf equation (2.32). As for an ergodic process (2.132) yields: K and 1 N C1 !p
K !1
1 N C1
!R
K !1
(2.143)
(2.144)
We nd that the LS solution tends toward the Wiener solution for sufciently large K , that is cls
K !1
! copt
(2.145)
In other words, for K ! 1 the covariance method gives the same solution as the autocorrelation method. In scalar notation, (2.141) becomes a system of N equations in N unknowns:
N 1 X i D0
n D 0; 1; : : : ; N
(2.146)
2.3.1
From (2.128), taking the gradient with respect to cn , we have rcn E D rcn ;I E C jrcn ;Q E D
K 1 X kDN 1
[ x .k
n/e.k/
x.k j x.k
C j . j x .k D 2
K 1 X kDN 1
n/e.k/ x .k
n/e.k/
If we denote with femin .k/g the estimation error found with the optimum coefcient values, cls , then the optimum coefcients must satisfy the conditions
K 1 X kDN 1
emin .k/x .k
n/ D 0
n D 0; 1; : : : ; N
(2.148)
152
which represent the time-average version of the statistical orthogonality principle (2.36). Moreover, being y.k/ a linear combination of fx.k n/g, n D 0; 1; : : : ; N 1, we have
K 1 X kDN 1
(2.149)
Equation (2.149) expresses the fundamental result: the optimum lter output sequence is orthogonal to the minimum estimation error sequence.
jy.k/j2 D c H c
(2.151)
observing (2.130). Note that for c D cls we have d.k/ D y.k/ C emin .k/ then, because of the orthogonality (2.149) between y and emin , it follows that Ed D E y C Emin from which, substituting (2.141) in (2.151), we get Emin D Ed
H where E y D cls .
(2.152)
(2.153)
Ey
(2.154)
from the denition (2.131) of #.n/ we get 2 3 2 #.0/ x .N 1/ x .N / 6 #.1/ 7 6 x .N 2/ x .N 1/ 6 7 6 6 7D6 : : : : : : 4 5 4 : : : .0/ .2/ x x #.N 1/
: : : x .K : : : x .K : :: : : : .K ::: x
(2.156)
153
that is D TH d Thus, using the (2.139) and (2.157), the normal equation (2.141) becomes T H Tcls D T H d (2.158) (2.157)
Associated with system (2.158), it is useful to introduce the system of equations for the minimization of E, Tc D d From (2.158), if .T H T/
1
(2.159)
TH d
(2.160)
TH d
(2.161)
We note how both formulae (2.160) and (2.161) depend only on the desired signal samples and input samples. Moreover, the solution c is unique only if the columns of T are linearly independent, that is the case of non-singular T H T. This requires at least K N C 1 > N , that is the system of equations (2.159) must be overdetermined with more equations than unknowns.
can be related to the input data matrix T as y D Tc This relation will still be valid for c D cls , and from (2.160) we get y D Tcls D T.T H T/
1
(2.163)
TH d
(2.164)
The matrix O D T.T H T/ 1 T H can be thought of as a projection operator dened on the space generated by the columns of T. Let I be the identity matrix: the difference O? D I ODI T.T H T/
1
TH
(2.166)
154
emin y
is the complementary projection operator, orthogonal to O . In fact, from (2.164) y D Od and from (2.165) emin D d y D O ?d (2.168) (2.167)
(2.169)
In Figure 2.11 an example illustrating the relation among d, y, and emin is given.
2.3.2
If the inverse of .T H T/ does not exist, the solution of the LS problem (2.160) must be reexamined. This is what we will do in this section after taking a closer look at the associated system of equations (2.159). In general, let us consider the solutions to a linear system of equations Tc D d with T N N square matrix. If T obtained in various ways [5]:
1
(2.170)
1d
1. If T is triangular and non-singular, a solution to the system (2.170) can be found by the successive substitutions method with O.N 2 / operations. 2. In general, if T is non-singular, one can use the Gauss method, which involves three steps: a. Factorization of T T D LU (2.171)
with L lower triangular having all ones along the diagonal and U upper triangular;
155
b. Solution of the system in z Lz D d through the successive substitutions method; c. Solution of the system in c Uc D z through the successive substitutions method. This method requires O.N 3 / operations and O.N 2 / memory locations. 3. If T is Hermitian and non-singular, the factorization (2.171) becomes the Cholesky decomposition: T D LL H (2.174) (2.173) (2.172)
with L lower triangular having non-zero elements on the diagonal. This method requires O.N 3 / operations, about half as many as the Gauss method. 4. If T is Toeplitz and non-singular, one can use the generalized Shur algorithm with a complexity of O.N 2 /: generally it is applicable to all T structured matrices [6]. We also recall the Kumar fast algorithm [4]. However, if T 1 does not exist, e.g., because T is not a square matrix, it is necessary to use alternative methods to solve the system (2.170) [5]: in particular we will consider the method of the pseudo-inverse. First, we will state the following result.
156
In (2.177) the fi g, i D 1; : : : ; R, are singular values of T. Being U and V unitary, it follows U H TV D as illustrated in Figure 2.12. Denition 2.1 The pseudo-inverse of T, L N , of rank R, is given by the matrix T# D V where
# #
(2.180)
UH D
R X i D1
i 1 vi uiH
(2.181)
D 1 0 0 0
D diag 1 1 ; 2 1 ; : : : ; R 1
(2.182)
We nd an expression of T# for the two cases in which T has full rank,4 that is R D min.L ; N /. Case of an overdetermined system (L > N ) and R D N . Note that in this case the system (2.170) has more equations than unknowns. Using the above relations it can be shown that T# D .T H T/
1
TH
(2.183)
In this case T# d coincides with the solution of system (2.141). Case of an underdetermined system (L < N ) and R D L. Note that in this case there are fewer equations than unknowns, hence there are innite solutions to the system (2.170). Again, it can be shown that T# D T H .TT H /
1
(2.184)
157
(2.185)
(2.186)
TH
(2.187)
i2
vi
(2.188)
(2.189)
i2
T H ui
(2.190)
Only solutions (2.185) in the cases (2.186) and (2.187) coincide with the solution (2.142). The computation of the pseudo-inverse T# directly from SVD and the expansion of c in terms of fui g, fvi g and fi2 g have two advantages with respect to the direct computation of T# in the form (2.187), for L > N and rank.T/ D N , or in the form (2.189), for L < N and rank.T/ D L:
158
1. The SVD also gives the rank of T through the number of non-zero singular values. 2. The required accuracy in computing T# via SVD is almost halved with respect to the computation of .T H T/ 1 or .TT H / 1 . There are two algorithms to determine the SVD of T: the Jacobi algorithm and the Householder transformation [7]. We conclude citing two texts [8, 9], which report examples of realizations of the algorithms described in this section.
Bibliography
[1] S. Haykin, Adaptive lter theory. Englewood Cliffs, NJ: Prentice-Hall, 3rd ed., 1996. [2] M. L. Honig and D. G. Messerschmitt, Adaptive lters: structures, algorithms and applications. Boston, MA: Kluwer Academic Publishers, 1984. [3] P. Delsarte and Y. V. Genin, The split Levinson algorithm, IEEE Trans. on Acoustics, Speech and Signal Processing, vol. 34, pp. 470478, June 1986. [4] R. Kumar, A fast algorithm for solving a Toeplitz system of equations, IEEE Trans. on Acoustics, Speech and Signal Processing, vol. 33, pp. 254267, Feb. 1985. [5] G. H. Golub and C. F. van Loan, Matrix computations. Baltimore and London: The Johns Hopkins University Press, 2nd ed., 1989. [6] N. Al-Dhahir and J. M. Ciof, Fast computation of channel-estimate based equalizers in packet data transmission, IEEE Trans. on Signal Processing, vol. 43, pp. 2462 2473, Nov. 1995. [7] S. A. T. W. H. Press, B. P. Flannery and W. T. Vetterling, Numerical Recipes. New York: Cambridge University Press, 3rd ed., 1988. [8] L. S. Marple Jr., Digital spectral analysis with applications. Englewood Cliffs, NJ: Prentice-Hall, 1987. [9] S. M. Kay, Modern spectral estimation-theory and applications. Englewood Cliffs, NJ: Prentice-Hall, 1988. [10] S. M. Kay, Fundamentals of statistical signal processing: estimation theory. Englewood Cliffs, NJ: Prentice-Hall, 1993.
159
Appendix 2.A
MMSE estimation
Let pd ./ and px ./ be the probability density functions of d and x, respectively, and pdjx . j / the conditional probability density function of d given x D ; moreover let px ./ 6D 0, 8. We wish to determine the function h that minimizes the mean-square error, that is Z C1 Z C1 2 [ h./]2 pdx .; / d d J D E[e ] D
1 1
Z D
C1 1
px ./
C1
where the relation pdx .; / D px ./ pdjx . j / is used. Theorem 2.2 The estimator h./ that minimizes J is given by the expected value of d given x D , h./ D E[d j x D ] Proof. The integral (2.193) is minimum when the function Z C1 [ h./]2 pdjx . j / d
1
(2.194)
(2.195)
is minimized for every value of . Using the variational method (see Appendix 8.A), we nd that this occurs if Z C1 [ h./] pdjx . j / d D 0 8 (2.196) 2
1
160
pdjx . j / d D
C1 1
pdx .; / d px ./
(2.197)
from which the (2.194) follows. O An alternative to the MMSE criterion for determining d is given by the maximum a posteriori probability (MAP) criterion, which yields O d D arg max pdjx . j /
(2.198)
where the notation arg max is dened in (6.21). If the distribution of d is uniform, the MAP criterion becomes the maximum likelihood (ML) criterion, where O d D arg max pxjd . j /
(2.199)
Examples of both MAP and ML criteria are given in Chapters 6 and 14. Example 2.A.1 Let d and x be two jointly Gaussian r.v.s with mean values md and mx , respectively, and covariance c D E[.d md /.x mx /]. After several steps, it can be shown that [10] h./ D md C c . 2 x mx / (2.200)
c x
2 (2.201)
Example 2.A.2 Let x D d C w, where d and w are two statistically independent r.v.s. For w 2 N .0; 1/ and d 2 f 1; 1g with P[d D 1] D P[d D 1] D 1=2, it can be shown that h./ D tanh./ (2.202)
the estimation of d is obtained by applying the following theorem, whose proof is similar to the case of a single observation.
161
O d/2 ] is given by
(2.204)
Z D
In the following, to simplify the formulation we will refer to r.v.s with zero mean. Example 2.A.3 Let d, x D [x1 ; : : : ; x N ]T , be real-valued jointly Gaussian r.v.s with zero mean and the following second order description: Correlation matrix of observations R D E[x xT ] Cross-correlation vector p D E[dx] For x D , it can be shown that h./ D pT R and
2 Jmin D d 1
(2.205)
(2.206)
(2.207)
pT R
(2.208)
162
p
1
pT R
(2.211)
Note that the r.v.s are assumed to have zero mean. Observation 2.3 Comparing (2.210) and (2.211) with (2.207) and (2.208), respectively, we note that, in the case of jointly Gaussian r.v.s, linear estimation coincides with optimum MMSE estimation.
(2.212)
(2.213)
Rx D E[x xT ] Rd D E[d d ] The problem is to determine a linear transformation of x, given by O d D CT x C b O such that d is a close replica of d in the mean-square error sense.
T
(2.219)
Denition 2.3 The linear minimum mean-square error (LMMSE) estimator, consisting of the N M matrix C, and of the M 1 vector b, coincides with the linear function of the observations (2.219) that minimizes the cost function J D E[jjd O djj2 ] D
M X mD1
E[jdm
O dm j2 ]
(2.220)
In other words, the optimum coefcients C and b are the solution of the following problem: min J
C;b
(2.221)
163
We note that in the formulation of Section 2.1 we have xT D [x.k/; : : : ; x.k d D [d.k/]
T
N C 1/]
(2.222) (2.223)
that is M D 1, and the matrix C becomes a column vector. We determine now the expression of C and b in terms of the correlation matrices introduced above. First of all, we observe that if d and x have zero mean, then b D 0, Q Q Q since the choice of b D b 6D 0 implies an estimator CT x C b with a larger value of the cost function. In fact J D E[jjd D E[jjd D E[jjd Q CT x Q CT xjj2 ] Q bjj2 ] 2RefE[.d Q Q Q CT x/ H b]g C jjbjj2 (2.224)
Q Q CT xjj2 ] C jjbjj2
Q being E[d] D E[x] D 0. The (2.224) implies that the choice b D 0 yields the minimum value of J . Without loss of generality, we will assume that both x and d are zero mean random vectors. Scalar case. For M D 1, d D d1 , and
T O O d D d1 D c1 x D xT c1
(2.225)
with c1 column vector with N coefcients. In this case the problem (2.221) leads again to the Wiener lter; the solution is given by Rx c1 D rd1 x where rd1 x is dened by (2.214). O Vector case. For M > 1, d and d are M-dimensional vectors. Nevertheless, since the function (2.220) operates on single components, the vector problem (2.221) leads to M O O O scalar problems, each with input x and output d1 ; d2 ; : : : ; d M , respectively. Therefore the columns of the matrix C, cm , satisfy equations of the type (2.226), Rx cm D rdm x m D 1; : : : ; M (2.227) (2.226)
hence, based on the denition (2.215), it results in C D Rx 1 Rxd Thus, the optimum estimator in the LMMSE sense is given by O d D .Rx 1 Rxd /T x (2.229) (2.228)
164
Value of the cost function. On the basis of the estimation error eDd with correlation matrix Re D E[e eT ] D Rd Rdx C C H Rxd C C H Rx C (2.231) O d (2.230)
the cost function (2.220) is given by the trace of Re , J D tr[Re ] Substituting (2.228) in (2.232), yields Jmin D tr[Rd Rdx Rx 1 Rxd ] (2.233) (2.232)
Chapter 3
We reconsider the Wiener lter introduced in Section 2.1. Given two random processes x and d, we want to determine the coefcients of a FIR lter having input x, so that the lter output y is a replica as accurate as possible of the process d. Adopting, for example, the mean-square error criterion, it is required that the autocorrelation matrix R of the lter input vector, and the cross-correlation p between the desired output and the input vector be known (see (2.15)). Estimating these correlations is usually difcult.1 Moreover, the optimum solution requires solving a system of equations with a computational complexity that is at least proportional to the square of the number of lter coefcients. In this chapter, we develop iterative algorithms with low computational complexity to obtain an approximation of the Wiener solution. We will consider transversal FIR lters2 with N coefcients. In general the coefcients may vary with time. The lter structure at instant k is illustrated in Figure 3.1. We dene: 1. Coefcient vector at instant k: cT .k/ D [c0 .k/; c1 .k/; : : : ; c N 2. Input vector at instant k: xT .k/ D [x.k/; x.k The output signal is given by y.k/ D
N 1 X i D0 1 .k/]
(3.1)
1/; : : : ; x.k
N C 1/]
(3.2)
ci .k/x.k
i/ D xT .k/c.k/
(3.3)
Comparing y.k/ with the desired response d.k/, we obtain the estimation error3 e.k/ D d.k/ y.k/ (3.4)
1 2
Two estimation methods are presented in Section 1.11.2. For the analysis of IIR adaptive lters we refer the reader to [1, 2]. 3 In this chapter the denition of the estimation error is given as the difference between the desired signal and the lter output. Depending on the application, the estimation error may be dened using the opposite sign. Some caution is therefore necessary in using the equations of an adaptive lter.
166
x(k) Tc
x(k-1) Tc
x(k-2) Tc
x(k-N+1)
c0 (k)
c1 (k)
c 2(k)
c N-1 (k)
+
y(k) e(k) + d(k)
Depending on the cost function associated with fe.k/g, in Chapter 2 two classes of algorithms have been developed: 1. mean-square error (MSE), 2. least squares (LS). In the following sections we will present iterative algorithms for each of the two classes.
3.1
Assuming that x and d are individually and jointly WSS, analogously to (2.17), J .k/ can be written as
2 J .k/ D d
c H .k/p
p H c.k/ C c H .k/Rc.k/
(3.6)
where R and p are dened respectively in (2.16) and (2.15). The optimum WienerHopf solution is c.k/ D copt , where copt is given by (2.34). The corresponding minimum value of J .k/ is Jmin , given by (2.40).
3.1.1
Our rst step is to realize a deterministic iterative procedure to compute copt . We will see that this method avoids the computation of the inverse R 1 , however, it requires that R and p be known.
167
(3.7)
where rc.k/ J .k/ denotes the gradient of J .k/ with respect to c (see (2.18)), is the adaptation gain, a real-valued positive constant, and k is the iteration index, in general not necessarily coinciding with time instants. As J .k/ is a quadratic function of the vector of coefcients, from (2.31) we nd rc.k/ J .k/ D 2.Rc.k/ hence c.k C 1/ D c.k/ .Rc.k/ p/ (3.9) p/ (3.8)
In the scalar case (N D 1), for real-valued signals the above relations become: J .k/ D Jmin C rx .0/.c0 .k/ and rc0 J .k/ D @ J .k/ D 2rx .0/.c0 .k/ @c0 copt;0 / (3.11) copt;0 /2 (3.10)
The iterative algorithm is given by: c0 .k C 1/ D c0 .k/ rx .0/.c0 .k/ copt;0 / (3.12)
The behavior of J and the sign of rc0 J .k/ as a function of c0 is illustrated in Figure 3.2. In the two-dimensional case (N D 2), the trajectory of rc J .k/ is illustrated in Figure 3.3. We recall that in general the gradient vector for c D c.k/ is orthogonal to the locus of points with constant J that includes c.k/.
J
Jmin 0
c opt,0
c0(1) c0(0)
Figure 3.2. Behavior of J and sign of the gradient vector rc in the scalar case (N D 1).
0< 0= c0
0>
168
Figure 3.3. Loci of points with constant J and trajectory of rc J in the two-dimensional case (N D 2).
R]c.k/ C Rcopt
Starting at k D 0 with arbitrary c.0/, that is with 1c.0/ D c.0/ copt , we determine now the conditions for the convergence of c.k/ to copt or, equivalently, for the convergence of 1c.k C 1/ to 0. Using the decomposition (1.369), R D U U H , where U is the unitary matrix formed of eigenvectors of R, and is the diagonal matrix of eigenvalues fi g, i D 1; : : : ; N , and setting (see (2.46)) .k/ D U H 1c.k/ equation (3.15) becomes .k C 1/ D [I ].k/ (3.17) (3.16)
169
Hence, i .k/ as a function of k is given by (see Figure 3.4): i .k/ D .1 i /k i .0/ (3.19)
The i-th component of the vector .k/ converges, that is i .k/ if and only if j1 or, equivalently, 1<1 i < 1 (3.22) i j < 1 (3.21) !0
k!1
8i .0/
(3.20)
As i is positive (see (1.360)), we have that the algorithm converges if and only if 0<< 2 i i D 1; 2; : : : ; N (3.23)
If max (min ) is the largest (smallest) eigenvalue of R, observing (3.23) the convergence condition can be expressed as 0<<
(k) i
2 max
(3.24)
Figure 3.4. Plot of i .k/ as a function of k for i < 1 and j1 i j < 1. In the case i > 1 and j1 i j < 1, i .k/ is still decreasing in magnitude, but it assumes alternating positive and negative values.
170
Correspondingly, observing (3.16) and (3.19) we obtain the expression of the vector of coefcients, given by c.k/ D copt C U.k/ D copt C [u1 ; u2 ; : : : ; u N ].k/ D copt C
N X i D1 N X i D1
ui i .k/
(3.25)
D copt C
ui .1
i /k i .0/
u i;n .1
i /k i .0/
n D 0; : : : ; N
(3.26)
where u i;n is the n-th component of ui . In (3.26) the term u i;n .1 i /k i .0/ characterizes the i-th mode of convergence. Note that, if the convergence condition (3.23) is satised, each coefcient cn , n D 0; : : : ; N 1, converges to the optimum solution as a weighted sum of N exponentials, each with the time constant4 i D 1 ln j1 i j ' 1 i i D 1; 2; : : : ; N (3.27)
i j
(3.28)
(3.29)
As illustrated in Figure 3.5, the solution is obtained for 1 min D .1 max / (3.30)
The time constant is the number of iterations needed to reduce the signal associated with the i-th mode of a factor e.
171
|1- i | 1
()
0 0
1 opt 1 max i
1 min
Figure 3.5. Plot of and j1 i j as a function of for different values of i : min < i < max .
from which we get opt D and .opt / D 1 2 max min min D D max C min max C min .R/ 1 .R/ C 1 (3.32) max 2 C min (3.31)
where .R/ D max =min is the eigenvalue spread (1.376). We note that other values of (associated with max or min ) cause a slower mode. We emphasize that .opt / is a monotonic function of the eigenvalue spread (see Figure 3.6), and consequently the larger the eigenvalue spread the slower the convergence.
i ji .k/j2
(3.33)
i .1
i /2k i2 .0/
(3.34)
Consequently, if the condition for convergence is veried, for k ! 1, J .k/ ! Jmin as a weighted sum of exponentials. The i-th mode will have a time constant given by MSE;i D 1 2 ln j1 i j ' 1 2i (3.35)
172
0.8
(opt)
0.6
0.4
0.2
5 (R)
10
Figure 3.6. .opt / as a function of the eigenvalue spread .R/ D max =min .
assuming i 1. We note that (3.34) is different from (3.26) because of the presence of i as weight of the i-ith mode: consequently, modes associated with small eigenvalues tend to weigh less in the convergence of J .k/. In particular, let us examine the two-dimensional case (N D 2). Recalling the observation that J .k/ increases more rapidly (slowly) in the direction of the eigenvector corresponding to D max ( D min ) (see Figure 2.4), we have the following two cases. Case 1 for 1 2 . Choosing c.0/ on the 2 axis (in correspondence of max ) we have the following situations: 8 >< 1 > > > max > > > > < 1 if D > max > > > > > > >> 1 : max the iterative algorithm has a non-oscillatory behavior the iterative algorithm converges in one iteration the iterative algorithm has a trajectory that oscillates around the minimum (3.36)
Let umin and umax be the eigenvectors corresponding to min and max , respectively. If no further information is given regarding the initial condition c.0/, choosing D opt the algorithm exhibits monotonic convergence along umin , and an oscillatory behavior around the minimum along umax .
173
Case 2 for 2 D 1 . Choosing D 1=max the algorithm converges in one iteration, independently of the initial condition c.0/.
3.1.2
The LMS or stochastic gradient algorithm is an algorithm with low computational complexity that provides an approximation to the optimum WienerHopf solution without requiring knowledge of R and p. Actually, the following instantaneous estimates are used: O R.k/ D x .k/xT .k/ and O p.k/ D d.k/x .k/ The gradient vector (3.8) is thus estimated to be5 rc.k/ J .k/ D D D 2d.k/x .k/ C 2x .k/xT .k/c.k/ 2x .k/[d.k/ 2x .k/e.k/ xT .k/c.k/] (3.39) (3.38) (3.37)
The equation for the adaptation of the lter coefcients, where k now denotes a given instant, becomes c.k C 1/ D c.k/ that is c.k C 1/ D c.k/ C e.k/x .k/ (3.41)
1 2 rc.k/ J .k/
(3.40)
Observation 3.1 The same equation is obtained for a cost function equal to je.k/j2 , whose gradient is given by rc.k/ J .k/ D 2e.k/x .k/ (3.42)
Implementation
The block diagram for the implementation of the LMS algorithm is shown in Figure 3.7, with reference to the following parameters and equations.
We note that (3.39) represents an unbiased estimate of the gradient (3.8), but in general it also exhibits a large variance.
174
x(k)
Tc *
x(k-1)
Tc *
x(k-2)
x(k-N+1)
c0 (k) ACC
c1 (k) ACC
c2 (k) ACC
e(k) d(k)
y(k) +
Figure 3.7. Block diagram of an adaptive transversal lter adapted according to the LMS algorithm.
Parameters. Required parameters are: 1. N , number of coefcients of the lter. 2. 0 < < Filter. 2 . statistical power of input vector
Adaptation. 1. Estimation error e.k/ D d.k/ 2. Coefcient vector adaptation c.k C 1/ D c.k/ C e.k/x .k/ Initialization. If no a priori information is available, we set c.0/ D 0 (3.46) (3.45) y.k/ (3.44)
The accumulators (ACC) in Figure 3.7 are used to memorize coefcients, which are updated by the current value of e.k/x .k/.
175
Computational complexity
For every iteration we have 2N C 1 complex multiplication (N due to ltering and N C 1 to adaptation) and 2N complex additions. Therefore the LMS algorithm has a complexity of O.N /.
Canonical model
The LMS algorithm operates with complex-valued signals and coefcients. We can rewrite complex-valued quantities as follows. Input vector Desired signal x.k/ D x I .k/ C jx Q .k/ d.k/ D d I .k/ C jd Q .k/ (3.47) (3.48) (3.49) (3.50) (3.51)
Coefcient vector c.k/ D c I .k/ C jc Q .k/ Output lter y.k/ D y I .k/ C j y Q .k/
Using the above denitions and considering separately real and imaginary terms, we derive the new equations: y I .k/ D xT .k/c I .k/ I e I .k/ D d I .k/ e Q .k/ D d Q .k/ c Q .k C 1/ D c Q .k/ y I .k/ y Q .k/ [e I .k/x Q .k/ e Q .k/x I .k/] xT .k/c Q .k/ Q (3.52) (3.53) (3.54) (3.55) (3.56) (3.57)
Therefore a complex-valued LMS algorithm is equivalent to a set of two real-valued LMS algorithms with cross-coupling. This scheme is adopted in practice if only processors that use real arithmetic are available.
! copt
(3.58) (3.59)
!0
176
x(k)
0
2
0
50
100
150
c(k)
0.5
1
0
E[c(k)] a
50
100
150
2
0
J(k)=E[ |e(k)|2 ]
2
4
6
8
10
12
J
0
min
50
100
150
Figure 3.8. Realizations of input fx.k/g, coefcient fc.k/g and squared error fje.k/j2 g for a one-coefcient predictor (N D 1), adapted according to the LMS algorithm.
In other words, it is required that the mean of the iterative solution converges to the Wiener Hopf solution and the mean of the estimation error approaches zero. To show the weakness of this criterion, in Figure 3.8 we illustrate the results of a simple experiment for an input x given by a real-valued AR(1) random process: 2 (3.60) x.k/ D a x.k 1/ C w.k/ w.k/ 2 N 0; w
2 2 where a D 0:95, and x D 1 (i.e. w D 0:097). For a rst-order predictor we adapt the coefcient, c.k/, according to the LMS algorithm with D 0:1. For c.0/ D 0 and k 0, we compute
1. Predictor output y.k/ D c.k/x.k 2. Prediction error e.k/ D d.k/ 3. Coefcient update c.k C 1/ D c.k/ C e.k/x.k 1/ (3.63) y.k/ D x.k/ y.k/ (3.62) 1/ (3.61)
177
In Figure 3.8, realizations of the processes fx.k/g, fc.k/g and fje.k/j2 g are illustrated, as well as mean values E[c.k/] and J .k/ D E[je.k/j2 ], estimated by averaging over 500 realizations; copt and Jmin represent the WienerHopf solution. From the plots in Figure 3.8 we observe two facts: 1. The random processes x and c exhibit a completely different behavior, for which they may be considered uncorrelated. It is interesting to observe that this hypothesis corresponds to assuming the lter input vectors, fx.k/g, statistically independent. Actually, for small values of , c follows mean statistical parameters associated with the process x and not the process itself. 2. Convergence of the mean is an easily reachable objective. By itself, however, it does not yield the desired results, because the iterative solution c may exhibit very large oscillations around the optimum solution. A constraint on the amplitude of the oscillations must be introduced. Convergence in the mean-square sense. E[jjc.k/ copt jj2 ] ! constant
k!1 k!1
J .k/ D E[je.k/j2 ]
! J .1/
In other words, at convergence, both the mean of the coefcient error vector norm and the output mean-square error must be nite. The quantity J .1/ Jmin D Jex .1/ is the MSE in excess and represents the price paid for using a random adaptation algorithm for the coefcients rather than a deterministic one, such as the steepest-descent algorithm. In any case, we will see that the ratio Jex .1/=Jmin can be made small by choosing a small adaptation gain . We note that the coefcients are obtained by averaging in time the quantity e.k/x .k/. Choosing a small the adaptation will be slow and the effect of the gradient noise on the coefcients will be strongly attenuated.
3.1.3
We recall the following denitions. 1. Coefcient error vector 1c.k/ D c.k/ 2. Optimum lter output error emin .k/ D d.k/ We also make the following assumptions. xT .k/copt (3.67) copt (3.66)
178
1. c.k/ is statistically independent of x.k/. 2. The components of the coefcient error vector, transformed according to U H , .k/ D [1 .k/; : : : ; N .k/] D U H 1c.k/ (see (3.16)), are orthogonal:
E[i .k/n .k/] D 0
i 6D n
i; n D 1; : : : ; N
(3.68)
This assumption is justied by the observation that the linear transformation that orthogonalizes both x.k/ (see (1.368)) and 1c.k/ (see (3.16)) in the gradient algorithm is given by U H . 3. Fourth-order moments can be expressed as products of second-order moments (see (3.97)). The adaptation equation of the LMS algorithm (3.41) can thus be written as 1c.k C 1/ D 1c.k/ C x .k/[d.k/ xT .k/c.k/] (3.69)
Adding and subtracting xT .k/copt to the terms within parentheses we obtain 1c.k C 1/ D 1c.k/ C x .k/[emin .k/ D [I xT .k/1c.k/] (3.70)
We note that 1c.k/ depends only on the terms x.k 1/; x.k 2/; : : : Moreover, with the change of variables (3.16), observing (3.33), the cost function6 J .k/ D E [je.k/j2 ]
x;c
(3.71)
i E[ji .k/j2 ]
(3.72)
As E[x .k/xT .k/] D R and the second term on the right-hand side of (3.73) vanishes for the orthogonality property (2.36) of the optimum lter, we obtain the same equation as in the case of the steepest descent algorithm: E[1c.k C 1/] D [I R]E[1c.k/] (3.74)
x;c
179
Consequently, for the LMS algorithm the convergence of the mean is obtained if 2 (3.75) max Observing (3.25) and (3.32), and choosing the value of D 2=.max C min /, the vector E[1c.k/] is reduced at each iteration at least by the factor .max min / = .max C min /. We can therefore assume that E[1c.k/] becomes rapidly negligible with respect to the mean-square error during the process of convergence. 0<<
x 2 .k//1c.k/x.k/emin .k/]
where the last term vanishes, as 1c.k/ has zero mean and is statistically independent of all other terms. Assuming7 moreover, E[x 4 .k/] D E[x 2 .k/x 2 .k/] D r2 .0/, (3.77) becomes x E[1c2 .k C 1/] D .1 C 2 r2 .0/ x Let D 1 C 2 r2 .0/ x 2rx .0/ (3.79) whose behavior as a function of is given in Figure 3.9. Then for the convergence of the difference equation (3.78) it must be j j < 1. Consequently must satisfy the condition 0<< Moreover, assuming rx .0/ 1, we get E[1c2 .1/] D D ' 2 rx .0/ Jmin rx .0/.2 rx .0// .2 Jmin rx .0// (3.81) 2 rx .0/ (3.80) 2rx .0//E[1c2 .k/] C 2 rx .0/Jmin (3.78)
Jmin 2
In other texts the Gaussian assumption is made, whereby E[x 4 .k/] D 3r2 .0/. The conclusions of the analysis x are similar.
180
0.8
0.6
0.4
0.2
0
1/ r (0) x
2/ r (0)
x
as a function of .
1c.k/x.k/
(3.82)
(3.83)
that is J .k/ D Jmin C rx .0/E[1c2 .k/] In particular, for k ! 1, we have J .1/ ' Jmin C rx .0/ The relative MSE deviation, or misadjustment, is: MSD D J .1/ Jmin Jex .1/ D D rx .0/ Jmin Jmin 2 (3.86) Jmin 2 (3.85) (3.84)
181
Let us dene Q x.k/ D [x1 .k/; : : : ; x N .k/]T D UT x.k/ Q Q and Q D x .k/Q T .k/ D U H x .k/xT .k/U x N N matrix with elements .i; n/ D xi .k/xn .k/ Q Q From (1.368) we get E[ ] D hence the components fxi .k/g are mutually orthogonal. Then (3.87) becomes Q .k C 1/ D [I ].k/ C emin .k/Q .k/ x (3.92) (3.91) i; n D 1; : : : ; N (3.90) (3.89) (3.88)
Recalling Assumption 1 of the convergence analysis, and assuming emin .k/ and x.k/ are not only orthogonal but also statistically independent, and consequently emin .k/ independent Q of x.k/, the correlation matrix of at the instant k C 1 can be expressed as E[ .k C 1/ T .k C 1/] D E[.I
/ .k/ T .k/.I
/] C 2 Jmin E[
(3.93)
Observing (3.91), the second term on the right-hand side of (3.93) is equal to 2 Jmin . Q Moreover, considering that .k/ is statistically independent of x.k/ and x.k/, the rst term can be written as
x;
E [.I
/ .k/ T .k/.I
/] D E [.I
x
/E ] .k/ T .k/[.I
/] (3.94)
D E[.I
/.I
Recalling Assumption 2 of the convergence analysis, we nd that the matrix E[ .k/ T .k/] is diagonal, with elements on the main diagonal given by the vector T .k/ D [1 .k/; : : : ; N .k/] D [E[j1 .k/j2 ]; : : : ; E[j N .k/j2 ]] (3.95)
Observing (3.90), the elements with indices .i; i/ of the matrix expressed by (3.94) are given by " # N X E i .k/ C 2 .i; n/ .n; i/n .k/ 2 .i; i/i .k/
nD1 N X nD1
D i .k/ C 2
2i i .k/
(3.96)
D i .k/ C 2
N X nD1
i n n .k/
2i i .k/
182
(3.97)
N N symmetric positive denite matrix with positive elements. From (3.93) and (3.96), we obtain the relation .k C 1/ D B.k/ C 2 Jmin Using the properties of B, the general decomposition (2.175) becomes B D V diag.1 ; : : : ; N /V H (3.101) (3.100)
where fi g denote the eigenvalues of B, and V is the unitary matrix formed by the eigenvectors fvi g of B. After simple steps, similar to those applied to get (3.25) from (3.13), and using the relation N rx .0/ D tr[R] D
N X i D1
(3.102)
the solution of the vector difference equation (3.100) is given by: .k/ D
N X i D1
Ki ik vi C
Jmin 1 N rx .0/
k0
(3.103)
where 1 D [1; 1; : : : ; 1]T . In (3.103) the constants fKi g are determined by the initial conditions Jmin i D 1; : : : ; N (3.104) 1 Ki D viH .0/ 2 N rx .0/ where the components of .0/ depend on the choice of c.0/ according to (3.95) and (3.16):
H n .0/ D E[jn .0/j2 ] D E[jun 1c.0/j2 ]
n D 1; : : : ; N
(3.105)
Using (3.95) and (3.98), the cost function J .k/ given by (3.72) becomes J .k/ D Jmin C T .k/ (3.106)
183
Ci ik C
2 Jmin N rx .0/
(3.107)
The rst term on the right-hand side of (3.107) describes the convergence behavior of the mean-square error, whereas the second term gives the steady-state value. Therefore further investigation of the properties of the matrix B will allow us to characterize the transient behavior of J .
Basic results
From the above convergence analysis, we will obtain some fundamental properties of the LMS algorithm. 1. The transient behavior of J does not exhibit oscillations; this result is obtained by observing the properties of the eigenvalues of B. 2. The LMS algorithm converges if the adaptation gain satises the condition 0<< 2 statistical power of input vector (3.109)
In fact the adaptive system is stable and J converges to a constant steady-state value under the conditions ji j < 1, i D 1; : : : ; N . This happens if 0<< 2 N rx .0/ (3.110)
Conversely, if satises (3.110), from (3.99) the sum of the elements of the i-th row of B satises
N X nD1
[B]i;n D 1
i .2
N rx .0// < 1
(3.111)
A matrix with these properties and whose elements are all positive has eigenvalues with absolute value less than one. In particular, being
N X i D1
i D tr[R] D N rx .0/ D
N 1 X i D0
(3.112) i/j ]
2
E[jx.k
184
We recall that, for convergence of the mean, it must be 0<< but since
N X i D1
2 max
(3.113)
i > max
(3.114)
the condition for convergence in the mean-square implies convergence of the mean. In other words, convergence in the mean-square imposes a tighter bound to allowable values of the adaptation gain than that imposed by convergence of the mean (3.113). The new bound depends on the number of coefcients, rather than on the eigenvalue distribution of the matrix R. The relation (3.110) can be intuitively explained noting that, for a given value of , an increase in the number of coefcients causes an increase in the excess mean-square error due to uctuations of the coefcients around the mean value. Increasing the number of coefcients without reducing the value of leads to instability of the adaptive system. 3. Equation (3.107) reveals a simple relation between the adaptation gain and the value J .k/ in the steady state (k ! 1): J .1/ D 2 2 Jmin N rx .0/ N rx .0/ Jmin 2 N rx .0/ (3.115)
from which the excess MSE is given by Jex .1/ D J .1/ Jmin D (3.116)
and the misadjustment has the expression MSD D for 2=.N rx .0//. Jex .1/ N rx .0/ ' N rx .0/ D Jmin 2 N rx .0/ 2 (3.117)
Observations
1. For ! 0 all eigenvalues of B tend toward 1. 2. As shown below, a small eigenvalue of the matrix R (i ! 0) determines a large time constant for one of the convergence modes of J , as i ! 1. However, a large time constant of one of the modes implies a low probability that the corresponding term contributes signicantly to the mean-square error. Proof. If i D 0, the i-th row of B becomes .0; : : : ; 0; [B]i;i D 1; 0; : : : ; 0/. Consequently i D 1 and viT D .0; : : : ; 0; vi;i D 1; 0; : : : ; 0/. As T vi D 0, from (3.108) we get Ci D 0.
185
It is generally correct to state that a large eigenvalue spread of R determines a slow convergence of J to the steady state. However, the fact that modes with a large time constant usually contribute to J less than the modes that converge more rapidly, mitigates this effect. Therefore the convergence of J is less inuenced by the eigenvalue spread of R than would be the convergence of 1c.k/. 3. If all eigenvalues of the matrix R are equal,8 i D rx .0/, i D 1; : : : ; N , the maximum eigenvalue of the matrix B is given by imax D 1 rx .0/.2 N rx .0// (3.118)
The remaining eigenvalues of B do not inuence the transient behavior of J , since Ci D 0, i 6D i max . Proof. It is easily veried that imax is an eigenvalue of B and viT D N 1=2 max [1; 1; : : : ; 1] is the corresponding eigenvector. Moreover, the PerronFrobenius theorem afrms that the maximum eigenvalue of a positive matrix B is a positive real number and that the elements of the corresponding eigenvector are positive real numbers [4]. Since all elements of vimax are positive, it follows that imax is the maximum eigenvalue of B. Moreover, because vimax is parallel to T , the other eigenvectors of B are orthogonal to . Hence Ci D 0, i 6D i max . 4. If all eigenvalues of the matrix R are equal, i D rx .0/, i D 1; : : : ; N , combining (3.107) with the (3.118) and considering a time varying adaptation gain .k/, we obtain J .k C 1/ ' [1 .k/rx .0/.2 .k/N rx .0//]J .k/ C 2.k/rx .0/Jmin (3.119)
The maximum convergence rate of J is obtained for the adaptation gain opt .k/ D J .k/ Jmin 1 N rx .0/ J .k/ (3.120)
As the condition J .k/ Jmin is normally veried at the beginning of the iteration process, it results opt .k/ ' and J .k C 1/ ' 1 1 N J .k/ (3.122) 1 N rx .0/ (3.121)
186
We note that the number of iterations required to reduce the value of J .k/ by one order of magnitude is approximately 2:3N . 5. Thus (3.103) indicates that at steady state all elements of become equal. Consequently, recalling Assumption 2 of the convergence analysis, in steady state the lter coefcients are uncorrelated random variables with equal variance. The mean corresponds to the optimum vector copt . 6. In case the LMS algorithm is used to estimate the coefcients of a system that slowly changes in time, the adaptation gain has a lower bound larger than 0. In this case, the value of J in steady state varies with time and is given by the sum of three terms: Jtot .k/ D Jmin .k/ C Jex .1/ C J` (3.123)
where Jmin .k/ corresponds to the WienerHopf solution, Jex .1/ depends instead on the LMS algorithm and is directly proportional to , and J` depends on the ability of the LMS algorithm to track the system variations and expresses the lag error in the estimate of the coefcients. It turns out that J` is inversely proportional to . Therefore, for time varying systems must be chosen as a compromise between Jex and J` and cannot be arbitrarily small [5, 6, 7].
Final remarks
1. The LMS algorithm is easy to implement. 2. The relatively slow convergence is inuenced by , the number of coefcients and the eigenvalues of R. In particular it must be 0<< 2 2 D N rx .0/ statistical power of input vector (3.124)
3. Choosing a small results in a slow adaptation, and in a small excess MSE at convergence Jex .1/. For a large , instead, the adaptation is fast at the expense of a large Jex .1/. 4. Jex .1/ is determined by the large eigenvalues of R, whereas the speed of convergence of E[c.k/] is imposed by min . If the eigenvalue spread of R increases, the convergence of E[c.k/] becomes slower; on the other hand the convergence of J .k/ is less sensitive to this parameter. Note, however, that the convergence behavior depends on the initial condition c.0/ [8].
3.1.4
In the previous section, the basic version of the LMS algorithm was presented. We now give a brief introduction to other versions that can be used for various applications.
187
Leaky LMS
The leaky LMS algorithm is a variant of the LMS algorithm that uses the following adaptation equation: c.k C 1/ D .1 /c.k/ C e.k/x .k/ (3.125)
with 0 < rx .0/. This equation corresponds to the following cost function: J .k/ D E[je.k/j2 ] C E[jjc.k/jj2 ] where, as usual, e.k/ D d.k/ cT .k/x.k/ (3.127) (3.126)
In other words, the cost function includes an additional term proportional to the norm of the vector of coefcients. In steady state we get
k!1
lim E[c.k/] D .R C I/
(3.128)
It is interesting to give another interpretation to what has been stated. Observing (3.128), the application of the leaky LMS algorithm results in the addition of a small constant to the terms on the main diagonal of the correlation matrix of the input process; one obtains the same result by summing white noise with statistical power to the input process. Both approaches are useful to make irreversible an ill-conditioned matrix R, or to accelerate the convergence of the LMS algorithm. It is usually sufcient to choose two or three orders of magnitude smaller than rx .0/, in order not to modify substantially the original WienerHopf solution. Therefore the leaky LMS algorithm is used in cases where the Wiener problem is ill-conditioned, and multiple solutions exist.
Sign algorithm
There are adaptation equations that are simpler to implement, at the expense of a lower speed of convergence, for the same J .1/. Three versions are:9 8 > sgn.e.k//x .k/ < c.k C 1/ D c.k/ C e.k/ sgn.x .k// (3.130) > : sgn.e.k// sgn.x .k// Note that the rst version has as objective the minimization of the cost function J .k/ D E[je.k/j] (3.131)
The sign of a vector of complex-valued elements is dened as follows: sgn.x.k// D [sgn.x I .k// C j sgn.x Q .k//; : : : ; sgn.x I .k N C1// C j sgn.x Q .k N C1//] (3.129)
188
Sigmoidal algorithm
As extensions of the algorithms given in (3.128), the following adaptation equations may be considered: 8 > '.e.k//x .k/ < c.k C 1/ D c.k/ C e.k/'.x .k// (3.132) > : .k// '.e.k//'.x where '.a/ is the sigmoidal function (see Figure 3.10) [9]: '.a/ D tanh a 2 D 1 e 1Ce
a a
(3.133)
where is a positive parameter. There also exists a piecewise linear version of the sigmoidal function dened as 1 a '.a/ D > A > : 1 where A is a positive parameter. 8 > > < per a < per A (3.134)
Aa A
per a > A
1
0.8
0.6
0.4
0.2
(a)
0
0.2
0.4
0.6
0.8
1 1
0.8
0.6
0.4
0.2
0 a
0.2
0.4
0.6
0.8
189
Normalized LMS
In the LMS algorithm, if some x.k/ assume large values, the adaptation algorithm is affected by strong noise in the gradient. This problem can be overcome by choosing an adaptation gain of the type: D where 0 < < 2, and Q O Mx .k/ D jjxjj2 D or, alternatively, O O Mx .k/ D N Mx .k/ (3.137)
N 1 X i D0
Q O p C Mx .k/
(3.135)
jx.k
i/j2
(3.136)
O where Mx .k/ is the estimate of the statistical power of x.k/. A simple estimate is obtained by the iterative equation (see (1.468)): O O Mx .k/ D a Mx .k 1/ C .1 a/jx.k/j2 (3.138)
for a ' 1. In (3.135), p is a positive parameter that is introduced to avoid the denominator becoming too small; typically p' 1 Mx 10 (3.140)
The normalized LMS algorithm has a speed of convergence that is potentially higher than the standard algorithm, for uncorrelated as well as correlated input signals [10]. To be able to apply the normalized algorithm, however, some knowledge of the input process is necessary, in order to assign the values of Mx and p so that the adaptation process does not become unstable.
190
the behavior of J will be illustrated in Figure 3.11. 2. Decreasing . For a time-invariant system, the adaptation gain usually selected for application with the sign algorithm (3.130) is given by .k/ D 1 2 C k k0 (3.142)
with limited to the range [min ; max ]. Typical values are 1 ' 1 and 2 1. 4. Vector of values of . Let T D [0 ; : : : ; N
1 ];
a. Initially larger values i are chosen in correspondence of those coefcients ci that have larger amplitude.
J(k)
2 = 1 / 2 Jmin
191
b. i changes with time following the rule 8 if the i-th component of the gradient has al> i .k/ 1 > < ways changed sign in the last m 0 iterations i .k C 1/ D > > if the i-th component of the gradient has : i .k/ never changed sign in the last m 1 iterations (3.144) with limited to the range [min ; max ]. Typical values are m 0 ; m 1 2 f1; 3g and D 2.
3.1.5
We consider a real AR(2) process of unit power, described by the equation x.k/ D a1 x.k 1/ a2 x.k 2/ C w.k/ (3.145)
with w additive white Gaussian noise (AWGN), and a1 D 1:3 a2 D 0:995 (3.146)
From (1.547), the roots of A.z/ are given by %e '0 , where p % D a2 D 0:997 and '0 D cos
1
(3.147)
a1 2%
2 Being rx .0/ D x D 1, from the (1.552) we nd that the statistical power of w is given by 2 w D
1 a2 [.1 C a2 /2 1 C a2
a2 ] D 0:0057 D 1
22:4 dB
(3.149)
We construct a predictor for x of order N D 2 with coefcients cT D [c1 ; c2 ], as illustrated in Figure 3.12, using the LMS algorithm and some of its variants [13]. From (2.83) we expect to nd in steady state c' a (3.150)
192
that is c1 '
a1 , c2 '
1/ 1/ C c2 .k/ x.k 2/
(3.151)
e.k/ D x.k/
y.k/
(3.152)
For the predictor of Figure 3.12 we now consider various versions of the adaptive LMS algorithm and their relative performance. Example 3.1.1 (Standard LMS) The equation for updating the coefcient vector is c.k C 1/ D c.k/ C e.k/x.k 1/ (3.153)
Convergence curves are plotted in Figure 3.13 for a single realization and for the mean (estimated over 500 realizations) of the coefcients and of the squared prediction error, for D 0:04. In Figure 3.14 a comparison is made between the curves of convergence of the mean for three values of . We observe that, by decreasing , the excess error decreases, thus giving a more accurate solution, but the convergence time increases. Example 3.1.2 (Leaky LMS) The equation for updating the coefcient vector is c.k C 1/ D .1 /c.k/ C e.k/x.k 1/ (3.154)
Convergence curves are plotted in Figure 3.15 for a single realization and for the mean (estimated over 500 realizations) of the coefcients and of the squared prediction error,
193
Figure 3.13. Convergence curves for the predictor of order N D 2, obtained by the standard LMS algorithm.
a
0.2
1.2
1
0.8
0.2
c1(k)
0.6
0.4
0.2
0.2
c2(k)
0.4
0.6
0.8
a 1 2
200
400
600
800
1000
200
400
600
800
1000
J(k) (dB)
10
15
20
25
100
200
300
400
500
600
700
800
900
1000
Figure 3.14. Comparison among curves of convergence of the mean obtained by the standard LMS algorithm for three values of .
194
Figure 3.15. Convergence curves for the predictor of order N D 2, obtained by the leaky LMS.
for D 0:04 and D 0:01. We note that the steady-state values are worse than in the previous case. Example 3.1.3 (Normalized LMS ) The equation for updating the coefcient vector is c.k C 1/ D c.k/ C .k/e.k/x.k The adaptation gain is of the type .k/ D where x .k/ D a x .k O2 O2 1/ C .1 a/jx.k/j2 k0 (3.157) (3.158) Q p C N x .k/ O2 (3.156) 1/ (3.155)
D 0:97
(3.159)
(3.160)
195
Figure 3.16. Convergence curves for the predictor of order N D 2, obtained by the normalized LMS algorithm.
Convergence curves are plotted in Figure 3.16 for a single realization and for the mean (estimated over 500 realizations) of the coefcients and of the squared prediction error, for D 0:08. We note that, with respect to the standard LMS algorithm, the convergence is Q considerably faster. A direct comparison of the convergence curves obtained in the previous examples is given in Figure 3.17. Example 3.1.4 (Sign LMS algorithm) We consider the three versions of the sign LMS algorithm: (1) c.k C 1/ D c.k/ C sgn.e.k//x.k (2) c.k C 1/ D c.k/ C e.k/ sgn.x.k 1/, 1//, 1//.
A comparison of convergence curves is given in Figure 3.18 for the three versions of the sign LMS algorithm, for D 0:04. It turns out that version (2), where the estimation error in the adaptation equation is not quantized, yields the best performance in steady state. Version (3), however, yields fastest convergence. To decrease the prediction error in steady state for versions (1) and (3), the value of could be further lowered, at the expense of reducing the speed of convergence.
196
Figure 3.17. Comparison of convergence curves for the predictor of order N D 2, obtained by three versions of the LMS algorithm.
a1
1.2
0.2
0.2
0.8
c1(k)
0.6
0.4
0.2
0.2
c2(k)
0.4
0.6
0.8
a 1 2
200
400
600
800
1000
200
400
600
800
1000
ver.3
5
J(k) (dB)
ver.1
10
ver.2
15
20
25
100
200
300
400
500
600
700
800
900
1000
Figure 3.18. Comparison of convergence curves obtained by three versions of the sign LMS algorithm.
197
Observation 3.2 As observed on page 97, for an AR process x, if the order of the predictor is greater than the required minimum, the correlation matrix result is ill-conditioned with a large eigenvalue spread. Thus the convergence of the LMS prediction algorithm can be extremely slow and can lead to a solution quite different from the YuleWalker solution. In this case it is necessary to adopt a method that ensures the stability of the error prediction lter, such as the leaky LMS.
3.2
We now consider a recursive algorithm to estimate the vector of coefcients c by an LS method, named recursive least squares (RLS) algorithm. The RLS algorithm is characterized by a speed of convergence that can be one order of magnitude faster than the LMS algorithm, obtained at the expense of a larger computational complexity. With reference to the system illustrated in Figure 3.19, we introduce the following quantities: 1. Input vector at instant i xT .i/ D [x.i/; x.i 2. Coefcient vector at instant k cT .k/ D [c0 .k/; c1 .k/; : : : ; c N
1 .k/]
1/; : : : ; x.i
N C 1/]
(3.161)
(3.162)
3. Filter output signal at instant i, obtained for the vector of coefcients c.k/ y.i/ D cT .k/x.i/ D xT .i/c.k/
x(i) Tc x(i-1) Tc x(i-2) Tc x(i-N+1)
(3.163)
c0 (k)
c1 (k)
c2 (k)
cN-1 (k)
+ y(i) e(i)
+ d(i)
Figure 3.19. Reference system for a RLS adaptive algorithm.
198
4. Desired output at instant i d.i/ At instant k, based on the observation of the sequences fx.i/g fd.i/g i D 1; 2; : : : ; k (3.165) (3.164)
the criterion for the optimization of the vector of coefcients c.k/ is the minimum sum of squared errors up to instant k. Dening E.k/ D we want to nd min E.k/
c.k/ k X i D1
k i je.i/j2
(3.166)
(3.167)
xT .i/c.k/.
is a forgetting factor, that enables proper ltering operations even with nonstationary signals or slowly time-varying systems. The memory of the algorithm is approximately 1=.1 /. This problem is the classical LS problem (2.128), applied to a sequence of prewindowed samples with the exponential weighting factor k .
Normal equation
Using the gradient method, the optimum value of c.k/ satises the normal equation .k/c.k/ D .k/ where .k/ D
k X i D1
(3.168)
k i x .i/xT .i/
(3.169)
k X i D1
k i d.i/x .i/
(3.170)
.k/ .k/
(3.171)
199
(3.172)
We now recall the following identity known as matrix inversion lemma [12]. Let ADB
1
C CD
CH
(3.175)
DB
BC.D C C H BC/
CH B
(3.176)
D .k
1/
C D x .k/
DD1
(3.177)
.k/ D
.k
1/
1 .k
1 .k
1/
1 C xT .k/ 1
1/x .k/
(3.178)
.k/
(3.179)
1/x .k/
1 .k
1 xT .k/
1/x .k/
(3.180)
also called the Kalman vector gain. From (3.178) we have the recursive relation P.k/ D
1
P.k
1/
k .k/xT .k/P.k
1/
(3.181)
200
We derive now a simpler expression for k .k/. From (3.180) we obtain k .k/[1 C from which we get k .k/ D D [ Using (3.181), it follows k .k/ D P.k/x .k/ Using the (3.174), the recursive equation to update the estimate of c is given by c.k/ D
1 1 1 T
x .k/
.k
1/x .k/] D
.k
1/x .k/
(3.182)
P.k P.k
1/x .k/ 1/
k .k/xT .k/P.k
k .k/xT .k/P.k
1/]x .k/
(3.184)
Substituting the recursive expression for P.k/ in the rst term, we get P.k 1/ .k 1/ 1/
1
k .k/xT .k/P.k
1/] .k 1/ .k 1/]
1/ C k .k/[d.k/
where in the last step (3.184) has been used. Dening the a priori estimation error, .k/ D d.k/ xT .k/c.k 1/ (3.187)
we note that xT .k/c.k 1/ is the lter output at instant k obtained by using the old coefcient estimate. In other words, from the a posteriori estimation error e.k/ D d.k/ xT .k/c.k/ (3.188)
we could say that .k/ is an approximated value of e.k/, that is computed before updating c. In any case the relation holds c.k/ D c.k 1/ C k .k/.k/ (3.189)
In summary, the RLS algorithm consists of four equations: k .k/ D P.k 1/x .k/ C xT .k/P.k 1/x .k/ xT .k/c.k 1/ (3.190) (3.191) (3.192) 1/ (3.193)
k .k/xT .k/P.k
201
In (3.190), k.k/ is the input vector ltered by P.k 1/ and normalized by the C xT .k/ P.k 1/x .k/. The term xT .k/P.k 1/x .k/ may be interpreted as the energy of the ltered input.
.k/ in (3.194)
k i x .i/xT .i/ C k I
with 1
This is equivalent to having for k 0 an all zero input with the exception of x. N C 1/ D . N C1 /1=2 . Consequently P.0/ D Typically
1 1
rx .0/
(3.196)
100 rx .0/
(3.197)
where rx .0/ is the statistical power of the input signal. In Table 3.1 we give a version of the RLS algorithm that exploits the fact that P.k/ (inverse of the Hermitian matrix .k/) is Hermitian, hence xT .k/P.k 1/ D [P.k 1/x .k/] H D T .k/ (3.198)
Initialization For k D 1; 2; : : : c.0/ D 0 P.0/ D 1 I 1/x .k/ 1 r.k/ D T .k/ .k/ Cx k .k/ D r.k/ .k/ .k/ D d.k/ xT .k/c.k 1/ c.k/ D c.k 1/ C .k/k .k/ P.k/ D
1 .P.k
.k/ D P.k
1/
k .k/ T .k//
202
k i jd.i/j2 D Ed .k
1/ C jd.k/j2
(3.199)
From the general LS expression (2.150), Emin .k/ D Ed .k/ observing (3.174) and (3.192) we get Emin .k/ D Ed .k 1/ C jd.k/j2 1/ C xT .k/d .k/][c.k H .k 1/c.k 1/ 1/ H .k/k .k/.k/ 1/ C .k/k .k/] (3.201) H .k/c.k/ (3.200)
[ H .k D Ed .k 1/
d .k/xT .k/c.k
1/ C d .k/.k/
Using the expression (3.179), and recalling that H .k/k .k/ D H .k/ D[
1
Moreover from (3.184) and (3.171) it follows that H .k/k .k/ D c H .k/x .k/ D x H .k/c .k/ Then (3.201) becomes Emin .k/ D Emin .k D Emin .k 1/ C d .k/.k/ 1/ C .k/[d .k/ c H .k/x .k/.k/ .xT .k/c.k// ] (3.204) (3.203)
Finally, the recursive relation is given by Emin .k/ D Emin .k We note that, as Emin .k/ is real, we get .k/e .k/ D .k/e.k/ that is .k/e .k/ is a real scalar value. (3.206) 1/ C .k/e .k/ (3.205)
203
N C 1/ output samples, the direct method (3.171) requires instead CCDIR D N 2 C N C K N3 N C1 (3.209)
We note that, if K N , the direct method is more convenient. In any case the RLS solution has other advantages: 1. It can be numerically more stable than the direct method. 2. It provides an estimate of the coefcients at each step and not only at the end of the data sequence. 3. For < 1 and 1=.1 / much less than the time interval it takes for the input samples to change statistics, the algorithm is capable of tracking the changes.
204
Figure 3.20. Convergence curves for the predictor of order N D 2, obtained by the RLS algorithm.
3.3
As observed in the previous section, the RLS algorithm has the disadvantage of requiring .2N 2 C 4N / multiplications per iteration. Therefore we will list a few fast algorithms, whose computational complexity increases linearly with N , the number of dimensions of the coefcient vector c. 1. Algorithms for transversal lters. The fast Kalman algorithm has the same speed of convergence as the RLS, but with a computational complexity comparable to that of the LMS algorithm. Exploiting some properties of the correlation matrix .k/, Falconer and Ljung [14] have shown that the recursive equation (3.193) requires only 10.2N C 1/ multiplications. Ciof and Kailath [15], with their fast transversal lter (FTF), have further reduced the number of multiplications to 7.2N C 1/. The implementation of these algorithms still remains relatively simple; their weak point resides in the sensitivity of the operations to round off errors in the various coefcients and signals. As a consequence the fast algorithms may become numerically unstable. 2. Algorithms for lattice lters. There are versions of the RLS algorithm for lattice structures that in the literature are called recursive least squares lattice (LSL) that have, in addition to a lower computational complexity than the standard RLS form, strong and weak points similar to those already discussed in the case of the LMS algorithm for lattice structures [12, 16].
205
divisions 0 N2 C 4N C 3 4
3. Algorithms for lters based on systolic structures. A particular structure is the QR decomposition-based LSL. The name comes from the use of an orthogonal triangularization process, usually known as QR decomposition, that leads to a systolic-type structure with the following characteristics: high speed of convergence; numerical stability, owing to the QR decomposition and lattice structure; a very efcient and modular structure, which does not require the a priori knowledge of the lter order and is suitable for implementation in very largescale integration (VLSI) technology. For further study on the subject we refer the reader to [17, 18, 19, 20, 21, 22].
3.3.1
In practice the choice of an algorithm must be made bearing in mind some fundamental aspects: computational complexity; performance in terms of speed of convergence, error in steady state, and tracking capabilities under non-stationary conditions; robustness, that is good performance achieved in the presence of a large eigenvalue spread and nite-precision arithmetic [5, 23]. Regarding the computational complexity per output sample, a brief comparison among LMS, RLS and FTF is given in Table 3.2. Although the FTF method exhibits a lower computational complexity than the RLS method, its implementation is rather laborious, therefore it is rarely used.
3.4
In this section some algorithms are examined that transform the input signal, for example from the time to the frequency domain, before adaptive ltering. With respect to the LMS algorithm, this approach may exhibit: a) lower computational complexity, or b) improved
206
convergence properties of the adaptive process. We will rst consider some adaptive algorithms in the frequency domain that offer some advantages from the standpoint of computational complexity [24, 25, 26, 27].
3.4.1
The basic scheme includes a lter that performs the equivalent operation of a circular convolution in the frequency domain. As illustrated in Figure 3.21, the method operates over blocks of N samples. The instant at which a block is processed is k D n N , where n is an integer number. Each input block is transformed using the DFT (see Section 1.4). The samples of the transformed sequence are denoted by fX i .n N /g, i D 0; 1; : : : ; N 1. We indicate with fDi .n N /g and fYi .n N /g, i D 0; 1; : : : ; N 1, respectively, the DFT of the desired output and of the adaptive lter output. Dening E i .n N / D Di .n N / Yi .n N /, the LMS adaptation algorithm is expressed as: Ci ..n C 1/N / D Ci .n N / C E i .n N /X i .n N / i D 0; 1; : : : ; N 1 (3.210)
In the following, lower case letters will be used to indicate sequences in the time domain, while upper case letters will denote sequences in the frequency domain.
207
Table 3.3 Comparison between the computational complexity of the LMS algorithm via FFT and the standard LMS for various values of the lter length N.
N 16 64 1024
terms of complex multiplications is given by N -point FFT of N real samples D N N -point FFT + 2 2 N N N N D C log2 4 2 2 2
(3.211)
then the algorithm requires a number of complex multiplications per output sample equal to N 1 C1 (3.212) log2 CCLMS f D 3 4 2 using the fact that fYi g and fCi g, i D 0; 1; : : : ; N 1 are Hermitian sequences. As each complex multiplication requires four real multiplications, the complexity in terms of real multiplications per output sample becomes N 3 log2 C1 (3.213) CCLMS f D 4 4 2 We note that the complexity in terms of real multiplications per output sample of the standard LMS algorithm is CCLMSt D 2N C 1 ' 2N (3.214)
A comparison between the computational complexity of the LMS algorithm via FFT and the standard LMS algorithm is given in Table 3.3. We note that the advantage of the LMS algorithm via FFT is non negligible even for small values of N . However, as the product between DFTs of two time sequences is equivalent to a circular convolution, the direct application of the scheme of Figure 3.21 is appropriate only if the relation between y and x is a circular convolution rather than a linear convolution.
3.4.2
We consider a block LMS adaptive algorithm in the time domain, for blocks of N input samples. Let us dene:
208
1. input vector at instant k xT .k/ D [x.k/; x.k 2. coefcient vector at instant n N cT .n N / D [c0 .n N /; c1 .n N /; : : : ; c N 3. lter output signal at instant n N C i y.n N C i/ D cT .n N /x.n N C i/ 4. error at instant n N C i e.n N C i/ D d.n N C i/ y.n N C i/ i D 0; 1; : : : N 1 (3.218) (3.217)
1 .n N /]
1/; : : : ; x.k
N C 1/]
(3.215)
(3.216)
The equation for updating the coefcients according to the block LMS algorithm is given by c..n C 1/N / D c.n N / C
N 1 X i D0
e.n N C i/x .n N C i/
(3.219)
As in the case of the standard LMS algorithm, the updating term is the estimate of the gradient at instant n N , .n N /. The above equations can be efciently implemented in the frequency domain by the overlap-save technique (see (1.112)). Assuming L-point blocks, where for example L D 2N , we dene10 C0 T .n N / D DFT[cT .n N /; 0; : : : ; 0] | {z }
N zeros
X0 .n N / D diag DFT[x.n N |
N /; : : : ; x.n N {z
block n 1
and Y0 .n N / D X0 .n N /C0 .n N / then the lter output at instants k 2 y.n N / 6 y.n N C 1/ 6 y.n N / D 6 : : 4 : y.n N C N D n N ; n N C 1; : : : ; n N C N 3 7 7 7 D last N elements of DFT 5 1, is given by (3.222)
[Y0 .n N /]
(3.223)
1/
209
We give now the equations to update the coefcients in the frequency domain. Let us consider the m-th component of the gradient, [.n N /]m D
N 1 X i D0
e.n N C i/x .n N C i
m/
m D 0; 1; : : : ; N
(3.224)
This component is given by the correlation between the error sequence fe.k/g and input fx.k/g, which is also equal to the convolution between e.k/ and x . k/. Let E0 T .n N / D DFT[0; : : : ; 0; d.n N / | {z } |
N zeros
y.n N /; : : : ; d.n N C N {z
1/
y.n N C N
errors in block n
1/] } (3.225)
[X0 .n N /E0 .n N /]
(3.226)
In the frequency domain, the adaptation equation (3.219) becomes .n N / C0 ..n C 1/N / D C0 .n N / C DFT 0
(3.227)
where 0 is the null vector with N elements. In summary, if 0 N N is the N N all zero matrix, I N N the N N identity matrix, and F the 2N 2N DFT matrix, then the following equations dene the fast LMS (FLMS): d0 T .n N /D[0T ; d.n N /; : : : ; d.n N C N 1/] 0 0 y0 .n N /D N N N N F 1 [X0 .n N /C0 .n N /] 0 N N I N N y0 .n N /] I 0 C0 ..n C1/N /DC0 .n N / C F N N N N F 0 N N 0 N N E0 .n N /DF[d0 .n N / (3.228) (3.229) (3.230)
1
[X0 .n N /E0 .n N /]
(3.231)
A comparison between the computational complexity of the FLMS algorithm and the standard LMS is given in Table 3.4.
210
211
N 16 32 64 1024
N R/E[c.n N /] C N p
where, as usual, R D E[x .k/xT .k/] and p D rdx D E[d.k/x .k/]. Recalling the analysis of the convergence of the steepest-descent algorithm of Section 3.1.1, we have
n!1
(3.234)
for 0 < < 2=.N max /, where max is the maximum eigenvalue of R. From these equations we can conclude: 1. The FLMS algorithm converges in the mean to the same solution of the LMS, however, must be smaller by a factor N in order to guarantee stability. 2. The time constant for the convergence of the i-th mode (for 1) is i D 1 i N blocks (3.235) samples
3. For 2=N max , it can be seen that the misadjustment is equal to that of the LMS algorithm: (3.236) MSD D tr[R] D N rx .0/ 2 2
3.5
We consider now some adaptive algorithms in the frequency domain that offer some advantages in terms of speed of convergence [28].
212
3.5.1
Basic scheme
Referring to Figure 3.23, we dene the following quantities. 1. Input vector at instant k xT .k/ D [x.k/; x.k 1/; : : : ; x.k N C 1/] (3.237)
with correlation matrix Rx D E[x .k/xT .k/]. 2. Transformed vector zT .k/ D [z 0 .k/; z 1 .k/; : : : ; z N In general, z.k/ D Gx.k/ where G is a unitary matrix of rank N : G 3. Coefcient vector at instant k cT .k/ D [c0 .k/; c1 .k/; : : : ; c N
1 .k/] 1 1 .k/]
(3.238)
(3.239)
D GH
(3.240)
(3.241)
213
4. Filter output signal y.k/ D zT .k/c.k/ D cT .k/z.k/ 5. Estimation error e.k/ D d.k/ y.k/ (3.243) (3.242)
6. Equation for updating the coefcients, LMS type: ci .k C 1/ D ci .k/ C i e.k/z i .k/ where i D Q E[jz i .k/j2 ] (3.245) i D 0; 1; : : : ; N 1 (3.244)
We note that each component of the adaptation gain vector has been normalized using the statistical power of the corresponding component of the transformed input vector. The various powers can be estimated, e.g., by considering a small window of input samples or recursively. Let
N
2 1 .k/j ]g
(3.246)
Then (3.244) can be written in vector notation as c.k C 1/ D c.k/ C e.k/ Q We nd that, for a suitable choice of , Q
k!1 1 N
z .k/
(3.247)
(3.248)
where Rz D E[z .k/zT .k/] D E[G x .k/xT .k/GT ] D G Rx GT and rdz D E[d.k/z .k/] D G E[d.k/x .k/] D G rdx Then copt D .G Rx G/ DG
1 1
(3.249)
(3.250)
G rdx
1
Rx 1 G
G rdx
D G H Rx 1 rdx D G H .Rx 1 rdx / where Rx 1 rdx is the optimum Wiener solution without transformation.
(3.251)
214
3.5.2
The convergence of the FLMS algorithm can be improved by dividing each component of the vector [X0 .n N /E0 .n N /] in (3.231) by the power of the respective component of X0 .n N /. Consequently the adaptation gain is adjusted to the various modes. This procedure, however, requires that the components of X0 .n N / are indeed uncorrelated.
3.5.3
(3.252)
x.k
m/e
mi j2 N
i D 0; 1; : : : ; N
(3.253)
The lters are of passband comb type, implemented by either 1) FFT with parallel input or 2) recursively with serial input to implement equations (3.254), as illustrated in Figure 3.24. In both cases the computational complexity to evaluate the output sample y.k/ is O.N log2 N /. Observation 3.3 A lter bank can be more effective in separating the various subchannels in frequency, even if more costly from the point of view of the computational complexity.
215
There are versions of the algorithm where each output z i .k/ is decimated, with the aim of reducing the number of operations. If fx.k/g and fd.k/g are real-valued signals, the lter coefcients satisfy the Hermitian property: ci .k/ D c N
1 i .k/
i D 0; 1; : : : ;
N 2
(3.255)
3.5.4
The LMS algorithm in the DCT domain is obtained by ltering the input by the lter bank of Figure 3.24, where the i-th lter has impulse response and transfer function given by, respectively, gi .k/ D cos and G i .z/ D Z[g i .k/] D .1 z 1
1 /.1
.2k C 1/i 2N
k D 0; 1; : : : ; N
(3.256)
. 1/i z z 2 cos N
N / cos 1
i 2N
2
(3.257)
Cz
216
Correspondingly, we have p N 1 2X x.k z 0 .k/ D N mD0 z i .k/ D X 2 N 1 x.k N mD0 m/ .2m C 1/i 2N i D0 (3.258)
m/ cos
i D 1; 2; : : : ; N
(3.259)
Ignoring the gain factor cos..=2N /i/, that can be included in the coefcient ci , even the ltering operation determined by G i .z/ can be implemented recursively [12]. We note that, if all the signals are real, the scheme can be implemented by using real arithmetic.
3.5.5
General observations
Orthogonalization algorithms are useful if the input has a large eigenvalue spread and fast adaptation is required. If the signals exhibit time-varying statistical parameters, usually these methods do not offer any advantage over the standard LMS algorithm. In general, they require larger computational complexity than the standard LMS.
3.6
Examples of application
We give now some examples of applications of the algorithms investigated in this chapter [1, 25, 29, 30].
3.6.1
System identication
We want to determine the relation between the input x and the output z of the system illustrated in Figure 3.25. We note that the observation d is affected by additive noise w, 2 having zero mean and variance w , assumed statistically independent of x.
Figure 3.25. System model in which we want to identify the relation between x and z.
217
Linear case
Assuming the system between z.k/ and x.k/ can be modelled as a FIR lter, the experiment illustrated in Figure 3.26 can be adopted to estimate the lter impulse response. Using an input x, known to both systems, we determine the output of the transversal lter c with N coefcients y.k/ D and the estimation error e.k/ D d.k/ The LMS adaptation equation follows, c.k C 1/ D c.k/ C e.k/x .k/ (3.262) y.k/ (3.261)
N 1 X i D0
ci .k/x.k
i/ D cT .k/x.k/
(3.260)
We analyze the specic case of an unknown linear FIR system whose impulse response has Nh coefcients. Assuming N Nh , we introduce the vector h with N components, hT D [h 0 ; h 1 ; : : : ; h Nh In this case, d.k/ D h 0 x.k/ C h 1 x.k D hT x.k/ C w.k/ 1/ C C h Nh
1 x.k 1 ; 0; : : : ; 0]
.Nh
(3.264)
For N Nh , and assuming the input x is white noise11 with statistical power rx .0/, we get R D E[x .k/xT .k/] D rx .0/I (3.265)
Figure 3.26. Adaptive scheme to estimate the impulse response of the unknown system.
218
and p D E[d.k/x .k/] D rx .0/h Then the WienerHopf solution to the system identication problem is given by copt D R and
2 Jmin D w 1
(3.266)
pDh
(3.267)
(3.268)
From (3.267) we see that the noise w does not affect the solution copt , consequently the expectation of (3.262) for k ! 1 (equal to copt ) is also not affected by w. Anyway, as seen in Section 3.1.3, the noise inuences the convergence process and the solution obtained by the adaptive LMS algorithm. The larger the power of the noise, the smaller must be so that c.k/ approaches E[c.k/]. In any case J .1/ 6D 0. On the other hand, if N < Nh then copt in (3.267) coincides with the rst N coefcients of h, and
2 Jmin D w C rx .0/ jj h.1/jj2
(3.269)
where
(3.270)
As the input x is white, the convergence behavior of the LMS algorithm (3.262) is easily determined. Let be dened as in (3.79): D 1 C rx .0/.2 N rx .0/ Let c.k/ D c.k/ copt ; then we get J .k/ D E[je.k/j2 ] D Jmin C rx .0/E[jj c.k/jj2 ] where E[jj c.k/jj2 ] D
k
2/
(3.271)
1 1
k0
(3.272)
The result (3.272) is obtained by (3.70) and the following assumptions: 1. c.k/ is statistically independent of x.k/;
2. emin .k/ is orthogonal to x.k/; 3. the approximation xT .k/ x .k/ ' N rx .0/ holds.
219
Indeed, (3.272) is an extension of (3.78). At convergence, for rx .0/ 1, it results in E[jj c.1/jj2 ] D and N J .1/ D Jmin 1 C rx .0/ 2 (3.274) N Jmin 2 (3.273)
A faster convergence and a more accurate estimate, for xed , are obtained by choosing a smaller value of N ; this, however, may increase the residual estimation error (3.269). Example 3.6.1 Consider an unknown system whose impulse response, given in Table 1.4 on page 26 as h 1 , has energy equal to 1.06. The noise is additive, white, and Gaussian with statistical power 2 w D 0:01. Identication via standard LMS and RLS adaptive algorithms is obtained using as input a maximal-length PN sequence of length L D 31 and unit power, Mx D 1. For a lter with N D 5 coefcients, the convergence curves of the mean-square error (estimated over 500 realizations) are shown in Figure 3.27. For the LMS algorithm, D 0:1 is chosen, which leads to a misadjustment equal to MSD D 0:26. As discussed in Appendix 3.B, as index of the estimate quality we adopt the ratio: 3n D
2 w E[jj hjj2 ]
(3.275)
Figure 3.27. Convergence curves of the mean-square error for system identication using LMS and RLS.
220
where h D c h is the estimate error vector. At convergence, that is for k D 30 in our example, it results: ( 3:9 for LMS (3.276) 3n D 7:8 for RLS We note that, even if the input signal is white, the RLS algorithm usually yields a better estimate than the LMS. However, for systems with a large noise power and/or slow timevarying impulse responses, the two methods tend to give the same performance in terms of speed of convergence and error in steady state. As a result it is usually preferable to adopt the LMS algorithm, as it leads to easier implementation.
where x.i/ 2 A, nite alphabet with M elements. Then z.k/ assumes values in an alphabet with at most M 3 values, which can be identied by a table or random-access memory (RAM) method, as illustrated in Figure 3.28. The cost function to be minimized is expressed as E[je.k/j2 ] D E[jd.k/ and the gradient estimate is given by rg je.k/j2 D O 2e.k/ (3.279) g.x.k//j2 ] O (3.278)
221
Therefore the LMS adaptation equation becomes g.x.k// D g.x.k// C e.k/ O O (3.280)
In other words, the input vector x.k/ identies a particular RAM location whose content is updated by adding a term proportional to the error. In the absence of noise, if the RAM is initialized to zero, the content of a memory location can be immediately identied by looking at the output. In practice, however, it is necessary to access each memory location several times to average out the noise. We note that, if the sequence fx.k/g is i.i.d., x.k/ selects in the average each RAM location the same number of times. An alternative method consists of setting y.k/ D 0 during the entire time interval devoted to system identication, and to update the RAM with the values of fd.k/g, according to the equation g.x.k// D g.x.k// C d.k/ O O k D 0; 1; : : : (3.281)
To complete the identication process, the value at each RAM location is scaled by the number of updates that have taken place for that location. This is equivalent to considering g.x/ D E[g.x/ C w] O (3.282)
We note that this method is a block version of the LMS algorithm with block length equal to the input sequence, where the RAM is initialized to zero, so that e.k/ D d.k/, and is given by the relative frequency of each address. Observation 3.4 In this section and in Appendix 3.B, the observation d and the input x are determined on the same time domain with sampling period Tc . Often, however, the input is determined on the domain with sampling period Tc , and the system output signal is determined on Tc =F0 . Using the polyphase representation (see Section 1.A.9) of d, it is convenient to represent the estimate of h determined on Tc =F0 as F0 estimates determined on Tc .
3.6.2
With reference to Figure 3.29, we consider two sensors: 1. Primary input, consisting of the desired signal s corrupted by additive noise w0 , d.k/ D s.k/ C w0 .k/ with s ? w0 (3.283)
2. Reference input, consisting of the noise signal w1 , with s ? w1 . We assume that w0 and w1 are in general correlated. w1 is ltered by an adaptive lter with coefcients fci g, i D 0; 1; : : : ; N lter output, given by y.k/ D
N 1 X i D0
1, so that the
ci .k/w1 .k
i/
(3.284)
222
is the most accurate replica of w0 .k/. Dening the error e.k/ D d.k/ y.k/ D s.k/ C w0 .k/ y.k/ (3.285) the cost function, assuming real-valued signals and recalling that s is orthogonal to the noise signals, is given by J D E[e2 .k/] D E[s 2 .k/] C E[.w0 .k/ We have two cases. 1. w1 and w0 are correlated: min J D rs .0/ C min E[.w0 .k/
c c
y.k//2 ]
(3.286)
y.k//2 ] D rs .0/
(3.287)
for y.k/ D w0 .k/. In this case e.k/ D s.k/. 2. w1 and w0 are uncorrelated: min J D E[.s.k/ C w0 .k//2 ] C min E[y 2 .k/]
c c
(3.288)
D E[.s.k/ C w0 .k//2 ] for y.k/ D 0. In this case e.k/ D d.k/ and the noise w0 is not cancelled.
General solution
With reference to Figure 3.30, for a general input x to the adaptive lter, the WienerHopf solution in the z-transform domain is given by (see (2.50)) Copt .z/ D Pdx .z/ Px .z/ (3.289)
223
Copt .z/ D
0 If w1 0, (3.290) becomes
(3.290)
Copt .z/ D
1 H .z/
(3.291)
224
3.6.3
Let
(3.292)
where s is the desired signal, and the sinusoidal term is the interferer. As shown in Figure 3.32, we take as reference signals x1 .k/ D B cos.2 f 0 kTc C '/ and x2 .k/ D B sin.2 f 0 kTc C '/ The adaptation equations of the LMS algorithm are c1 .k C 1/ D c1 .k/ C e.k/x 1 .k/ c2 .k C 1/ D c2 .k/ C e.k/x 2 .k/ (3.295) (3.296) (3.294) (3.293)
At convergence, the two coefcients c1 and c2 change the amplitude and phase of the reference signal to cancel the interfering tone. The relation between d and output e corresponds to a notch lter as illustrated in Figure 3.33. It is easy to see that x2 is obtained from x1 via a Hilbert lter (see Figure 1.28). We note that in this case x2 can be obtained as a delayed version of x1 .
3.6.4
With reference to Figure 3.34, the primary signal is a speech waveform affected by interference signals such as echoes and/or environmental disturbances. The reference signal
225
consists of a replica of the disturbances. At convergence, the adaptive lter output will attempt to subtract the interference signal, which is correlated to the reference signal, from the primary signal. The output signal is a replica of the speech waveform, obtained by removing to the best possible extent the disturbances from the input signal.
3.6.5
With reference to the simplied scheme of Figure 3.35, the speech signal of user A is transmitted over a transmission line consisting of a pair of wires (local loop) [31] to the central ofce A, where the signals in the two directions of transmission, i.e. the signal transmitted by user A and the signal received from user B, are separated by a device called
226
Figure 3.36. Conguration to remove the echo of signal A caused by the hybrid B.
hybrid. A similar situation takes place at the central ofce B, with the roles of the signals A and B reversed. Because of impedance mismatch, the hybrids give origin to echo signals that are added to the desired speech signals. For speech waveforms, the echo of signal A that is generated at the hybrid A can be ignored because it is not perceived by the human ear. The case for digital transmission is different, as will be discussed in Chapter 16. A method to remove echo signals is illustrated in Figure 3.36, where y is a replica of the echo. At convergence, e will consist of the speech signal B only.
3.6.6
In radio systems, to equalize the desired signal and remove interference, it is convenient to use several sensors, i.e. an antenna array, with the task of ltering signals in space, discriminating them through their angle of arrival. The signals of the array are then equalized to compensate for linear distortion introduced by the radio channel. A general scheme for wideband signals is illustrated in Figure 3.37. For narrowband signals, it is sufcient to substitute for each sensor the lter with a single complex-valued coefcient [32, 33] (see Section 8.18).
227
Figure 3.37. Antenna array to lter and equalize wideband radio signals.
3.6.7
For the cancellation of a periodic interfering signal, we can use the scheme of Figure 3.38, where: we note the absence of an external reference signal; the reference signal is generated by delaying the primary input; a delay 1 D DTc , where D is an integer, is needed to decorrelate the desired component of the primary signal from that of the reference signal, otherwise part of the desired signal would also be cancelled. On the other hand, to cancel a wideband interferer from a periodic signal it is sufcient to take the output of the adaptive lter (see Figure 3.39).
228
Figure 3.38. Scheme to remove a periodic interferer from a wideband desired signal.
Figure 3.39. Scheme to remove a wideband interferer from a periodic desired signal.
3. Bibliography
229
Note that in both schemes the adaptive lter acts as a predictor. Exploiting the general concept described above, an alternative scheme to that of Figure 3.32 is illustrated in Figure 3.40, where the knowledge of the frequency of the interfering signal is not required. In general, for D > 1 the scheme of Figure 3.40 requires many more than two coefcients, therefore it has a higher implementation complexity than that of the scheme of Figure 3.32. However, if the wideband signal can be modeled as white noise, then D D 1; hence, observing (1.555), for a sinusoidal interferer a secondorder predictor is sufcient.
Bibliography
[1] J. R. Treichler, C. R. Johnson Jr., and M. G. Larimore, Theory and design of adaptive lters. New York: John Wiley & Sons, 1987. [2] J. J. Shynk, Adaptive IIR ltering, IEEE ASSP Magazine, vol. 6, pp. 421, Apr. 1989. [3] G. Ungerboeck, Theory on the speed of convergence in adaptive equalizers for digital communication, IBM Journal of Research and Development, vol. 16, pp. 546555, Nov. 1972. [4] G. H. Golub and C. F. van Loan, Matrix computations. Baltimore and London: The Johns Hopkins University Press, 2nd ed., 1989. [5] S. H. Ardalan and S. T. Alexander, Fixed-point round-off error analysis of the exponentially windowed RLS algorithm for time varying systems, IEEE Trans. on Acoustics, Speech and Signal Processing, vol. 35, pp. 770783, June 1987. [6] E. Eweda, Comparison of RLS, LMS and sign algorithms for tracking randomly time varying channels, IEEE Trans. on Signal Processing, vol. 42, pp. 29372944, Nov. 1994. [7] W. A. Gardner, Nonstationary learning characteristics of the LMS algorithm, IEEE Trans. on Circuits and Systems, vol. 34, pp. 11991207, Oct. 1987. [8] V. Solo, The limiting behavior of LMS, IEEE Trans. on Acoustics, Speech and Signal Processing, vol. 37, pp. 19091922, Dec. 1989. [9] S. Haykin, Neural networks: a comprehensive foundation. New York: Macmillan Publishing Company, 1994. [10] S. C. Douglas, A family of normalized LMS algorithms, IEEE Signal Processing Letters, vol. 1, pp. 4951, Mar. 1994. [11] B. Porat and T. Kailath, Normalized lattice algorithms for least-squares FIR system identication, IEEE Trans. on Acoustics, Speech and Signal Processing, vol. 31, pp. 122128, Feb. 1983.
230
[12] S. Haykin, Adaptive lter theory. Englewood Cliffs, NJ: Prentice-Hall, 3rd ed., 1996. [13] J. R. Zeidler, Performance analysis of LMS adaptive prediction lters, IEEE Proceedings, vol. 78, pp. 17811806, Dec. 1990. [14] D. Falconer and L. Ljung, Application of fast Kalman estimation to adaptive equalization, IEEE Trans. on Communications, vol. 26, pp. 14391446, Oct. 1978. [15] J. M. Ciof and T. Kailath, Fast, recursive-least-squares transversal lter for adaptive ltering, IEEE Trans. on Acoustics, Speech and Signal Processing, vol. 32, pp. 304 337, Apr. 1984. [16] M. L. Honig and D. G. Messerschmitt, Adaptive lters: structures, algorithms and applications. Boston, MA: Kluwer Academic Publishers, 1984. [17] J. M. Ciof, High speed systolic implementation of fast QR adaptive lters, in Proc. ICASSP, pp. 15841588, 1988. [18] F. Ling, D. Manolakis, and J. G. Proakis, Numerically robust least-squares latticeladder algorithm with direct updating of the reection coefcients, IEEE Trans. on Acoustics, Speech and Signal Processing, vol. 34, pp. 837845, Aug. 1986. [19] P. A. Regalia, Numerical stability properties of a QR-based fast least squares algorithm, IEEE Trans. on Signal Processing, vol. 41, pp. 20962109, June 1993. [20] S. T. Alexander and A. L. Ghirnikar, A method for recursive least-squares ltering based upon an inverse QR decomposition, IEEE Trans. on Signal Processing, vol. 41, pp. 2030, Jan. 1993. [21] J. M. Ciof, The fast adaptive rotors RLS algorithm, IEEE Trans. on Acoustics, Speech and Signal Processing, vol. 38, pp. 631653, Apr. 1990. [22] Z.-S. Liu, QR methods of O.N / complexity in adaptive parameter estimation, IEEE Trans. on Signal Processing, vol. 43, pp. 720729, Mar. 1995. [23] J. A. Bucklew, T. G. Kurtz, and W. A. Sethares, Weak convergence and local stability properties of xed step size recursive algorithms, IEEE Trans. on Information Theory, vol. 39, pp. 966978, May 1993. [24] B. Widrow, Fundamental relations between the LMS algorithm and the DFT, IEEE Trans. on Circuits and Systems, vol. 34, pp. 814820, July 1987. [25] C. F. N. Cowan and P. M. Grant, Adaptive lters. Englewood Cliffs, NJ: Prentice-Hall, 1985. [26] N. J. Bershad and P. L. Feintuch, A normalized frequency domain LMS adaptive algorithm, IEEE Trans. on Acoustics, Speech and Signal Processing, vol. 34, pp. 452 461, June 1986. [27] R. P. Bitmead and B. D. O. Anderson, Adaptive frequency sampling lters, IEEE Trans. on Circuits and Systems, vol. 28, pp. 524543, June 1981.
3. Bibliography
231
[28] D. F. Marshall, W. K. Jenkins, and J. J. Murphy, The use of orthogonal transforms for improving performance of adaptive lters, IEEE Trans. on Circuits and Systems, vol. 36, pp. 474484, Apr. 1989. [29] B. Widrow and S. D. Stearns, Adaptive signal processing. Englewood Cliffs, NJ: Prentice-Hall, 1985. [30] O. Macchi, Adaptive processing: the LMS approach with applications in transmission. New York: John Wiley & Sons, 1995. [31] D. G. Messerschmitt, Echo cancellation in speech and data transmission, IEEE Journal on Selected Areas in Communications, vol. 2, pp. 283297, Mar. 1984. [32] I. J. Gupta and A. A. Ksienski, Adaptive antenna array for weak interfering signals, IEEE Trans. on Antennas Propag., vol. 34, pp. 420426, Mar. 1986. [33] L. L. Horowitz and K. D. Senne, Performance advantage of complex LMS for controlling narrow-band adaptive array, IEEE Trans. on Circuits and Systems, vol. 28, pp. 562576, June 1981. [34] D. G. Messerschmitt and E. A. Lee, Digital communication. Boston, MA: Kluwer Academic Publishers, 2nd ed., 1994. [35] S. W. Golomb, Shift register sequences. San Francisco: Holden-Day, 1967. [36] P. Fan and M. Darnell, Sequence design for communications applications. Taunton: Research Studies Press, 1996. [37] D. C. Chu, Polyphase codes with good periodic correlation properties, IEEE Trans. on Information Theory, vol. 18, pp. 531532, July 1972. [38] R. L. Frank and S. A. Zadoff, Phase shift pulse codes with good periodic correlation properties, IRE Trans. on Information Theory, vol. 8, pp. 381382, Oct. 1962. [39] A. Milewsky, Periodic sequences with optimal properties for channel estimation and fast start-up equalization, IBM Journal of Research and Development, vol. 27, pp. 426431, Sept. 1983. [40] R. L. Peterson, R. E. Ziemer, and D. E. Borth, Introduction to spread spectrum communications. Englewood Cliffs, NJ: Prentice-Hall, 1995. [41] R. Gold, Optimal binary sequences for spread spectrum multiplexing, IEEE Trans. on Information Theory, vol. 13, pp. 619621, Oct. 1967. [42] R. Gold, Maximal recursive sequences with 3-valued recursive cross-correlation functions, IEEE Trans. on Information Theory, vol. 14, pp. 154155, Jan. 1968. [43] S. L. Marple Jr., Efcient least squares FIR system identication, IEEE Trans. on Acoustics, Speech and Signal Processing, vol. 29, pp. 6273, Feb. 1981.
232
[44] J. I. Nagumo and A. Noda, A learning method for system identication, IEEE Trans. on Automatic Control, vol. 12, pp. 282287, June 1967. [45] S. N. Crozier, D. D. Falconer, and S. A. Mahmoud, Least sum of squared errors (LSSE) channel estimation, IEE Proceedings-F, vol. 138, pp. 371378, Aug. 1991. [46] N. Benvenuto, Distortion analysis on measuring the impulse response of a system using a cross-correlation method, AT&T Bell Laboratories Technical Journal, vol. 63, pp. 21712192, Dec. 1984.
3.A. PN sequences
233
Appendix 3.A
PN sequences
In this Appendix we introduce three classes of deterministic periodic sequences having spectral characteristics similar to those of a white noise signal, hence the name pseudonoise (PN) sequences.
Maximal-length sequences
Maximal-length sequences are binary PN sequences, also called r-sequences, that are generated recursively, e.g., using a shift-register (see page 877), and have period equal to L D 2r 1. Let f p.`/g, ` D 0; 1; : : : ; L 1, p.`/ 2 f0; 1g, be the values assumed by the sequence in a period. It can be shown that the maximal-length sequences enjoy the following properties [34, 35]. Every non-zero sequence of r bits appears exactly once in each period; therefore all binary sequences of r bits are generated, except the all zero sequence. The number of bits equal to 1 in a period is 2r 0 is 2r 1 1.
1,
A subsequence is intended here as a set of consecutive bits of the r-sequence. The relative frequency of any non-zero subsequence of length i r is 2r i '2 2r 1
i
(3.297)
and the relative frequency of a subsequence of length i < r with all bits equal to zero is 2r i 1 '2 2r 1
i
(3.298)
In both formulae the approximation is valid for a sufciently large r. The sum of two r-sequences, which are generated by the same shift-register, but with different initial conditions, is still an r-sequence. The linear span, that determines the predictability of a sequence, is equal to r [36]. In other words, the elements of a sequence can be determined by any 2r consecutive elements of the sequence itself, while the remaining elements can be produced by a recursive algorithm (see, e.g., the Berlekamp-Massey algorithm on page 891). A practical example is given in Figure 3.41 for a sequence with L D 15 (r D 4), which is generated by the recursive equation p.`/ D p.` 3/ p.` 4/ (3.299)
234
where denotes modulo 2 sum. Assuming initial conditions p. 1/ D p. 2/ D p. 3/ D p. 4/ D 1, applying (3.299) we obtain the sequence 1 0 0 |{z} |{z} 0 1 0 0 1 1 0 1 0 1 1 1 |{z} : : :
p.0/ p.1/ p.L 1/
(3.300)
Obviously, the all zero initial condition must be avoided. To generate sequences with a larger period L we refer to Table 3.5. The above properties make an r-sequence, even if deterministic and periodic, appear as a random i.i.d. sequence from the point of view of the relative frequency of subsequences of bits. It turns out that an r-sequence appears as random i.i.d. also from the point of view of the autocorrelation function. In fact, mapping 0 to 1 and 1 to C1, we get the following correlation properties. 1. Mean
L 1 1 X 1 p.`/ D L `D0 L
(3.301)
n/mod L D
8 < :
1 1 L
We note that, with the exception of the values assumed for .m/mod L D 0, the spectral density of maximal length sequences is constant.
3.A. PN sequences
235
Table 3.5 Recursive equations to generate PN sequences of length L D 2r 1, for different values of r.
r 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.` p.`/ D p.`
Period L D 2r 1/ 1/ p.` 2/ p.` 3/ p.` 3/ p.` 5/ p.` 6/ p.` 2/ p.` 5/ p.` 7/ p.` 9/ p.` 2/ p.` 1/ p.` 2/ p.` 14/ p.` 11/ p.` 14/ p.` 11/ p.` 14/ p.` 17/ p.`
2/ 3/ 4/ 5/ 6/ 7/ 3/ p.` 9/ 10/ 11/ 10/ p.` 11/ p.` 12/ p.` 15/ 13/ p.` 17/ 18/ 17/ p.` 20/ 18/ p.` 19/ 14/ p.` 16/ 11/ p.` 12/ p.` 13/ p.` 12/ 13/ 14/ 4/ p.` 8/
CAZAC sequences
The constant amplitude zero autocorrelation (CAZAC) sequences are complex-valued PN sequences with constant amplitude (assuming values on the unit circle) and autocorrelation function r p .n/ equal to zero for .n/mod L 6D 0. Because of these characteristics they are also called polyphase sequences [37, 38, 39]. Let L and M be two integer numbers that are relatively prime. The CAZAC sequences are dened as, for L even for L odd p.`/ D e j p.`/ D
M `2 L
` D 0; 1; : : : ; L ` D 0; 1; : : : ; L
1 1
(3.304) (3.305)
M `.`C1/ L ej
It can be shown that, in both cases, these sequences have the following properties.
236
1. Mean
L 1 1 X p.`/ D 0 L `D0
(3.306)
Gold sequences
In a large number of applications, as for example in spread-spectrum systems with codedivision multiple access (see Chapter 10), sets of sequences having one or both of the following properties [40] are required. Each sequence of the set must be easily distinguishable from its own time shifted versions. Each sequence of the set must be easily distinguishable from any other sequence of the set and from its time-shifted versions. An important class of periodic binary sequences that satisfy these properties, or, in other words, that have good autocorrelation and cross-correlation characteristics, is the set of Gold sequences [41, 42]. Construction of pairs of preferred r-sequences. In general the cross-correlation sequence (CCS) between two r-sequences may assume three, four or maybe even a greater number of values. We show now the construction of a pair of r-sequences, called preferred r-sequences [36], whose CCS assumes only three values. Let a D fa.`/g be an r-sequence with period L D 2r 1. We dene now another r-sequence of length L D 2r 1 obtained from the sequence a by decimation by a factor M, that is: b D fb.`/g D fa.M`/mod L g We make the following assumptions. rmod 4 6D 0, that is r must be odd or equal to odd multiples of 2, i.e. rmod 4 D 2. The factor M satises one of the following properties: M D 2k C 1 or M D 22k 2k C 1 k integer (3.310) (3.309)
3.A. PN sequences
237
For k determined as in the (3.310), dening g:c:d:.r; k/ as the greatest common divisor of r and k, let ( 1 r odd e D g:c:d:.r; k/ D (3.311) 2 rmod 4 D 2 Then the CCS between the two r-sequences a and b assumes only three values [35, 36]: rab .n/ D
L 1 1 X a.`/b .` L `D0
n/mod L
r Ce 2 2
1C2 1 1 2
r Ce 2
e 1
C2
e
2r
e 1
(3.312)
r Ce 2
r Ce 2 2
Example 3.A.1 (Construction of a pair of preferred r-sequences) Let the following r-sequence of period L D 25 1 D 31 be given: fa.`/g D .0000100101100111110001101110101/ (3.313)
As r D 5 and rmod 4 D 1, we take k D 1. Therefore e D g:c:d:.r; k/ D g:c:d:.5; 1/ D 1 and M D 2k C 1 D 21 C 1 D 3. The sequence fb.`/g obtained by decimation of the sequence fa.`/g is then given by fb.`/g D fa.3`/mod L g D .0001010110100001100100111110111/ The CCS between the two sequences, assuming 0 is mapped to 1, is: frab .n/g D 1 .7; 7; 1; 1; 1; 9; 7; 9; 7; 7; 1; 1; 7; 7; 1; 7; 1; 31 1; 9; 1; 1; 1; 1; 9; 1; 7; 1; 9; 9; 7; 1/ We note that, if we had chosen k D 2, then e D g:c:d:.5; 2/ D 1 and M D 22 C 1 D 5, or else M D 222 22 C 1 D 13. Construction of a set of Gold sequences. A set of Gold sequences can be constructed from any pair fa.`/g and fb.`/g of preferred r-sequences of period L D 2r 1. We dene the set of sequences: G.a; b/ D fa; b; a b; a Z b; a Z 2 b; : : : ; a Z L
1
(3.314)
(3.315)
bg
(3.316)
where Z is the shift operator that cyclically shifts a sequence to the left by a position. The set (3.316) contains L C 2 D 2r C 1 sequences of length L D 2r 1 and is called the set of Gold sequences. It can be proved [41, 42] that, for the two sequences fa 0 .`/g and fb0 .`/g
238
belonging to the set G.a; b/, the CCS as well as the ACS, with the exception of zero lag, assume only three values: 8 r C1 r C1 > 1C2 2 r odd 1 2 2 1< 1 ra 0 b0 .n/ D (3.317) r C2 r C2 L> : 1 1C2 2 rmod 4 D 2 1 2 2 Clearly, the ACS of a Gold sequence no longer has the characteristics of an r-sequence, as is seen in the next example. Example 3.A.2 (Gold sequence properties) Let r D 5, hence L D 25 1 D 31. From Example 3.A.1, the two sequences (3.313) and (3.314) are a pair of preferred r-sequences, from which it is possible to generate the whole set of Gold sequences. For example we calculate the ACS of fa.`/g and fb0 .`/g D fa.`/ b.` 2/g D a Z 2 b, and the CCS between fa.`/g and fb0 .`/g: fa.`/gD. 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1/ fb0 .`/gD. 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1; 1/ fra .n/gD (3.318)
(3.319)
frb0 .n/gD
frab0 .n/gD
1 . 1; 7; 7; 7; 1; 1; 1; 1; 1; 7; 9; 1; 1; 7; 1; 9; 7; 31 1 9; 7; 7; 9; 1; 7; 1; 9; 1; 1; 1; 9; 1/ (3.322)
239
Appendix 3.B
3.B.1
Correlation method
With reference to (3.264), which describes the relation between input and output of an unknown system with impulse response fh i g, i D 0; 1; : : : ; N 1, we take as an input signal white noise with statistical power rx .0/. To estimate the impulse response of a linear system, we observe that the cross-correlation between d and x is then proportional, with a factor rx .0/, to the impulse response fh i g. In fact, we have rdx .n/ D rzx .n/ D rx h.n/ D rx .0/h n (3.323)
In practice, instead of noise a PN sequence with period L, f p.i/g, i D 0; 1; : : : ; L 1, is used as input. We recall that the autocorrelation of a PN sequence is also periodic with period L and is given by (see Appendix 3.A): ( r p .n/ D1 '0 n D 0; L ; 2L ; : : : n D 1; : : : ; L 1; L C 1; : : :
(3.324)
Moreover, we recall that if the input to a time-invariant lter is periodic with period L, the output will also be periodic with period L. To estimate the impulse response fh i g, i D 0; 1; : : : ; N 1, we consider the scheme illustrated in Figure 3.42, where we choose L N , and an input sequence x with length of at least .L C 1/N samples, obtained by repeating f p.i/g; in other words, x.k/ D p.k/mod L . We assume a delay m 2 f0; 1; : : : ; N 1g, a rectangular window g Rc .k/ D .1=L/w L .k/, and that the system is started at instant k D 0. For k .N 1/ C .L 1/, the output v.k/ is given by
Figure 3.42. Correlation method to estimate the impulse response of an unknown system.
240
v.k/ D
L 1 X 1 u.k L `D0
`/ D
L 1 X 1 d.k L `D0
`/ p .k
m/mod L
L 1 X 1 D L `D0
"
N 1 X i D0
h i p.k `
i/mod L p .k #
m/mod L
Cw.k
N 1 X i D0
`/ p .k
L 1 1 X p.k L `D0
m/mod L
(3.325)
hi
i/mod L p .k `
m/mod L
C As
L 1 1 X p.k L `D0
L 1 1 X w.k L `D0
`/ p .k
m/mod L
i/mod L p .k
m/mod L D r p .m
i/mod L
(3.326)
h i r p .m
i/mod L C
L 1 1 X w.k L `D0
`/ p .k
m/mod L
(3.327)
If L 1, the second term on the right-hand side of (3.327) can be ignored, hence observing (3.324) we get v.k/ ' h m (3.328)
Mean and variance of the estimate of h m given by (3.327) are obtained as follows. 1. Mean E[v.k/] D assuming w has zero mean. 2. Variance
L 1 1 X var[v.k/] D var w.k L `D0 N 1 X i D0
h i r p .m
i/mod L
(3.329)
"
# `/ p .k
m/mod L
'
2 w L
(3.330)
241
Figure 3.43. Correlation method via correlator to estimate the impulse response of an unknown system.
Using the scheme of Figure 3.42, varying m from 0 to N 1 it is possible to get an estimate of the samples of the impulse response of the unknown system fh i g at the output of the lter g Rc . However, this scheme has two disadvantages: 1. it requires a very long computation time (N L); 2. it requires synchronization between the two PN sequences, at transmitter and receiver. Both problems can be resolved by memorizing, after a transient equal to N 1 instants, L consecutive output samples fd.k/g in a buffer and computing the correlation off-line: rdx .m/ D O 1 L
.N 1/C.L 1/ X kD.N 1/
d.k/ p .k
m/mod L ' h m
m D 0; 1; : : : ; N
(3.331)
An alternative scheme is represented in Figure 3.43: with steps analogous to those of the preceding scheme, we get v.k/ D
L 1 1 X d.k L `D0
.L
1/ C `/ p .` C .N 1/ C .L
1//mod L ' h .k
.N 1/ .L 1//mod L
(3.332)
After a transient of .N
O h i D v.i C .N
In other words, the samples at the correlator output from instant k D .N 1/ C .L 1/ to 2.N 1/ C .L 1/, give an estimate of the samples of the impulse response of the unknown system fh i g.
242
The quality of the estimate is measured by the signal-to-estimation error ratio 3e D jjhjj2 E[jj hjj2 ] (3.334)
On one hand, we have to take into consideration the noise present in the observed system and measured by (see Figure 3.42): 3D Mx jjhjj2 2 w (3.335)
where Mx is the statistical power of the input signal. In our case Mx D 1. Finally, we refer to the normalized ratio 3n D
2 w 3e D 3 Mx E[jj hjj2 ]
(3.336)
O We note that if we indicate with d.k/ the output of the identied system, O d.k/ D
N 1 X i D0
O h i x.k
i/
(3.337)
O O the fact that h 6D h causes d.k/ 6D z.k/, with an error given by z.k/ O d.k/ D
N 1 X i D0
.h i
O h i / x.k
i/
(3.338)
having variance Mx E[jj hjj2 ] for a white noise input. As a consequence, (3.336) measures the ratio between the variance of the additive noise of the observed system and the variance of the error at the output of the identied system. From (3.338) we note that the difference d.k/ O d.k/ D .z.k/ O d.k// C w.k/ (3.339)
consists of two terms, one due to the estimation error and the other due to the noise of the system.
3.B.2
x.k
n/mod L h n
(3.340)
where a PN sequence of period L D N , equal to the length of the impulse response to be estimated, is assumed as input signal x. Let us consider the vector zT D [z.k/;
243
z.k C 1/; : : : ; z.k C .L 1//], and a circulant matrix M whose rst row is [x.k/ mod L ; x.k 1/mod L ; : : : ; x.k .L 1//mod L ]. After an initial transient of L 1 samples, using the output samples fz.L 1/; : : : ; z.2.L 1//g we obtain a system of L linear equations in L unknowns, which in matrix notation can be written as z D Mh (3.341)
assuming k D L 1 in the denition of z and M. The system of equations (3.341) admits a unique solution if and only if the matrix M is non-singular. Because the input sequence is periodic, the system (3.341) can be solved very efciently, from a computational complexity point of view, in the frequency domain, rather than inverting the matrix M. Being M circulant, the product in (3.341) can be substituted by the circular convolution (see (1.105)) z.k/ D x
L
h.k/
kDL
1; : : : ; 2.L
1/
(3.342)
Letting Zm D DFT[z.k/], X m D DFT[x.k/], and Hm D DFT[h k ], (3.342) can be rewritten in terms of the discrete Fourier transforms as Zm D Xm Hm from which we get h k D DFT or, setting s.k/ D DFT
1 [1=X ], m 1
m D 0; : : : ; L
(3.343)
Zm Xm
k D 0; : : : ; L
(3.344)
hk D s
z.k/
(3.345)
d.k/ D s
z.k/ C s
w.k/ D h k C s
w.k/
(3.346)
2 Assuming that w is zero-mean white noise with power w , mean and variance of the estimate (3.346) are obtained as follows.
(3.347)
# O jh k O h k j2 D L E[jh k
2 h k j2 ] D L w
L 1 X i D0
js.i/j2
(3.348)
244
js.i/j2 D
L 1 L 1 1 X 1 X 1 jS j j2 D L jD0 L jD0 jX j j2
(3.349)
it is possible to particularize (3.336) for PN maximal-length and CAZAC sequences. In the rst case, from (3.303), it turns out X0 D 1 jX1 j2 D jX2 j2 D D jX L hence, observing (3.348), (3.336) becomes 3n D L C1 2L (3.351)
2 1j
(3.350) D L C1
For CAZAC sequences, from (3.308), we have that all terms jX j j2 are equal jX j j2 D L j D 0; 1; : : : ; L 1 (3.352)
2 and the minimum of (3.348) is equal to w , therefore 3n D 1. In other words, if L is large, CAZAC sequences yield 3 dB improvement with respect to the maximal-length sequences. Although this method is very simple, it has the disadvantage that, in the best case, it gives an estimate with variance equal to the noise variance of the original system.
3.B.3
The LS method
1/; : : : ; (3.353)
With reference to the system of Figure 3.42, letting xT .k/ D [x.k/; x.k x.k .N 1//], the noisy output of the unknown system can be written as d.k/ D hT x.k/ C w.k/ k D .N 1/; : : : ; .N 1/ C .L 1/
From (3.353) we see that the observation of L samples of the received signal requires the transmission of L T S D L C N 1 symbols of the training sequence fx.0/; x.1/; : : : ; x..N 1/ C .L 1//g. The unknown system can be identied using the LS criterion O [43, 44, 45]. For a certain estimate h of the unknown system, the sum of squared errors at the output is given by ED where, from (3.337), O O d.k/ D hT x.k/ As for the analysis of Section 2.3, we introduce the following quantities. (3.355)
N X 1 1CL kDN 1
jd.k/
O d.k/j2
(3.354)
245
jd.k/j2
(3.356)
i; n D 0; : : : ; N
(3.357)
x .k
i/ x.k
n/
(3.358)
1/]
(3.359)
d.k/ x .k
n/
(3.360)
As the matrix is determined by a suitably chosen training sequence, we can assume that is positive denite and therefore the inverse exists. The solution to the LS problem yields O hls D with a corresponding error equal to Emin D Ed O H hls (3.363)
1
(3.362)
1 in the (3.362) can be pre-computed and memorized, because We observe that the matrix it depends only on the training sequence. In some applications it is useful to estimate the O variance of the noise signal w that, observing (3.339), for h ' h can be assumed equal to
.w / D O2
1 Emin L
(3.364)
246
x.N
: : : 1/ C .L
1/
(3.365)
1/; : : : ; d..N
1/ C .L
1//]
(3.366)
D IHo
(3.367)
IHo
(3.368)
d.k/ x .k/
(3.369)
w.k/ x .k/
(3.370)
Consequently, substituting (3.371) in (3.362), the estimation error vector can be expressed as hD
1
(3.372)
247
2 If w is zero-mean white noise with variance w , is a zero-mean random vector with correlation matrix 2 R D E[ T ] D w
(3.373)
Therefore,
(3.374)
In particular,
2 E[jj hjj2 ] D w tr[.
(3.375)
]/
where I is the N N identity matrix. The elements on the diagonal of 1=L, and (3.376) yields 3n D L N
are equal to
(3.378)
The (3.378) gives a good indication of the relation between the number of observations L, the number of system coefcients N , and 3n . For example, doubling the length of the training sequence, 3n also doubles. Now, using as training sequence a maximal-length sequence of periodicity L, and indicating with 1 N N the matrix with all elements equal to 1, the correlation matrix can be written as D .L C 1/I From (3.379) the inverse is given by
1
1 N N
(3.379)
1 N N 1 IC D L C1 L C1 N
(3.380)
In Figure 3.44 the behavior of 3n is represented as a function of N , for CAZAC sequences (solid line) and for maximal-length sequences (dotted-dashed line), with parameter L. We make the following observations.
248
Figure 3.44. 3n vs. N for CAZAC sequences (solid line) and maximal-length sequences (dotted-dashed line), for various values of L.
For a given N , choosing L N , the two sequences yield approximately the same 3n . The worst case is obtained for L D N ; for example, for L D 15 the maximallength sequence yields a value of 3n that is about 3 dB lower than the upper bound (3.378). We note that the frequency method operates for L D N . For a given value of L, because of the presence of the noise w, the estimate of the coefcients becomes worse if the number of coefcients N is larger than the number of coefcients of the system Nh . On the other hand, if N is smaller than Nh , the estimation error may assume large values (see (3.270)). For sparse systems, where the number of coefcients may be large, but only a few of them are non-zero, the estimate is usually very noisy. Therefore, after obtaining the estimate, it is necessary to set to zero all coefcients whose amplitude is below a certain threshold. If the correlation method (3.331) is adopted, we get 1 O hD L where is given by (3.359). Observing (3.371), we get 1 1 hD I hC L L
(3.382)
249
Consequently the estimate is affected by a BIAS term equal to ..1=L/ I/h, and has a covariance matrix equal to .1=L 2 / R . In particular, using (3.373), it turns out E[jj hjj ] D and 3n D 1 tr[ ] C L2 1 1 L I h
2 2
1 L
I h
2 w tr[ ] L2
(3.383)
1 2 w
(3.384)
Using a CAZAC sequence, from (3.377) the second term of the denominator in (3.384) vanishes, and 3n is given by (3.378). In fact, for a CAZAC sequence, as 1 is diagonal, the LS method (3.362) coincides with (3.324) is strictly true and the correlation method (3.331). Using instead a maximal-length sequence, from (3.379) we get 1 L I h
2
1 jj.1 L2
I/ hjj2 D
X 1 N 1 jh i L 2 i D0
H.0/j2
(3.385)
D where H.0/ D PN
1 i D0
1 .jjhjj2 C .N L2
2/ jH.0/j2 /
where 3 is dened in (3.335). We observe that using the correlation method, we obtain the same values 3n (3.381) as the LS method, if L is large enough to satisfy the condition 3 C .N 2/ jH.0/j2 <L 2 w
3.B.4
We refer to the system model of Figure 3.42. Let us assume that w and h are statistically independent random processes, whose second-order statistic is known.
250
For a known input sequence x, we desire to estimate h using the LMMSE method given in the Appendix 2.A, from the observation of the noisy output sequence d, which now will be denoted as o. We note that in the LS method the observation was the transmitted signal x, while the desired signal was given by the system noisy output d. The observation is now given by d and the desired signal is the system impulse response h. Consequently, some caution is needed to apply (2.229) to the problem under investigation. Recalling the denition (3.366) of the observation vector o, and (2.229), the LMMSE estimator is given by O hLMMSE D .Ro 1 Roh /T o ; (3.388)
where we have assumed E[o] D 0 and E[h] D 0. Consequently, (3.388) provides an estimate only of the random (i.e. the non-deterministic) component of the channel impulse response. Now, from (3.353), letting wT D [w.N 1/; : : : ; w..N 1/ C .L 1//] (3.389)
be a random vector with noise components, we can write o D Ih C w Assuming that the sequence fx.k/g is known, we have Ro D E[o o T ] D I Rh I T C Rw and Roh D E[o hT ] D I Rh Then (3.388) becomes O hLMMSE D [.I Rh I T C Rw /
1
(3.390)
(3.391)
(3.392)
I Rh ] T o
(3.393)
Using the matrix inversion lemma (3.176), (3.393) can be rewritten as O hLMMSE D [.R / h
2 If Rw D w I, we have 2 O hLMMSE D [w .R / h 1 1
C I H .R / w
I]
I H .R / w
(3.394)
C I H I]
IHo
(3.395)
We note that with respect to the LS method (3.368), the LMMSE method (3.395) introduces a weighting of the components given by D I H o, which depends on the ratio between the noise variance and the variance of h. If the variance of the components of h is large, then Rh is also large and likely Rh 1 can be neglected in (3.395). We conclude by recalling that Rh is diagonal for a WSSUS radio channel model (see (4.221)), and the components of h are derived by the power delay prole.
251
For an analysis of the estimation error we can refer to (2.233), which uses the error O vector h D hLMMSE h having a correlation matrix R1h D f[.R / h
2 If Rw D w I, we get 2 2 R1h D w f[w .R / h 1 1
C I H .R / w
I] g
(3.396)
C I H I] g
(3.397)
Moreover, in general E[jj hjj2 ] D tr[R1h ] This result can be compared with that of the LS method given by (3.375). (3.398)
3.B.5
In the case of continuous-time systems, the scheme of Figure 3.42 can be modied to that of Figure 3.45 [46], where the noise is neglected. A PN sequence of period L, repeated several times, is used to modulate in amplitude the pulse t Tc =2 (3.399) g.t/ D wTc .t/ D rect Tc The modulated output signal x is therefore given by x.t/ D
C1 X i D0
p.i/mod L g.t
i Tc /
(3.400)
x./x .
t/d
(3.401)
Figure 3.45. Basic scheme to measure the impulse response of an unknown system.
252
As g has nite support of length Tc , rx .t/ has a simple expression given by rx .t/ D
L 1 1 X r p .`/rg .t Tc `D0
`Tc /
0 t L Tc
(3.402)
where, in the case of g.t/ given by (3.399), we have Z Tc g./g. t/d D Tc 1 rg .t/ D
0
(3.403)
Substituting (3.403) in (3.402) and assuming a maximal-length PN sequence, with r p .0/ D 1 and r p .`/ D 1=L for ` D 1; : : : ; L 1, we obtain t L Tc 1 jtj 1 rect jtj 1 C 1C (3.404) rx .t/ D L L Tc 2Tc 2 as shown in Figure 3.46 for L D 8. If the output z of the unknown system to be identied is multiplied by a delayed version of the input, x .t /, and the result is ltered by an ideal integrator between 0 and L Tc with impulse response 1 1 t L Tc =2 (3.405) w L Tc .t/ D rect g Rc .t/ D L Tc L Tc L Tc we obtain v.t/ D D 1 L Tc Z
t
u./d
t L Tc
/d
(3.406)
253
Figure 3.47. Sliding window method to measure the impulse response of an unknown system.
Therefore the output assumes a constant value v equal to the convolution between the unknown system h and the autocorrelation of x evaluated in . Assuming 1=Tc is larger than the maximum frequency of the spectral components of h, and L is sufciently large, the output v is approximately proportional to h. /. The scheme represented in Figure 3.47 is an alternative to that of Figure 3.45, of simpler implementation because it does not require synchronization of the two PN sequences at transmitter and receiver. In this latter scheme, the output z of the unknown system is multiplied by a PN sequence having the 0 same characteristics of the transmitted sequence, but a different clock frequency f 0 D 1=Tc0 , related to the clock frequency f 0 D 1=Tc of the transmitter by the relation 1 0 (3.407) f0 D f0 1 K where K is a parameter of the system. We consider the function Z L Tc 1 rx 0 x . / D [x 0 ./] x. / d L Tc 0
(3.408)
where is the delay at time t D 0 between the two sequences. As time elapses, the delay between the two sequence diminishes of the quantity .t=Tc0 /.Tc0 Tc / D t=K , so that Z t 1 t L Tc 0 for t L Tc [x ./] x. / d ' rx 0 x (3.409) L Tc t L Tc K If K is sufciently large, we can assume that rx 0 x . / ' rx . / At the output of the lter g Rc , given by (3.405), therefore we have Z t Z C1 1 [h. /x. / d ][x 0 ./] d v.t/ D L Tc t L Tc 0 Z C1 t L Tc 0x d D h. /rx K 0 Z C1 t L Tc ' d h. /rx K 0 (3.410)
(3.411)
254
C1
h. /rx .t 0
/ d
(3.412)
where the integral in (3.412) coincides with the integral in (3.406). If K is sufciently large (an increase of K clearly requires a greater precision and hence a greater cost of 0 the frequency synthesizer to generate f 0 ), it can be shown that the approximations in (3.409) and (3.410) are valid. Therefore the systems in Figure 3.47 and in Figure 3.45 are equivalent.
Chapter 4
Transmission media
The rst two sections of this chapter introduce several parameters that are associated with the electrical characteristics of electronic devices. The fundamental properties of various transmission media will be discussed in the remaining sections.
4.1
256
In the cascade of several 2-port networks, the frequency response G Ch . f / of each network is given by the ratio VL =V1 . Therefore, with reference to Figure 4.2c, we have GCh . f / D G1
o.
f /G L . f /
(4.1)
where G1 o . f / and G L . f / denote the frequency responses of the 2-port network and of the load, respectively. We note that in some cases the frequency response could be dened as VL =I1 , I L =V1 , or I L =I1 . For these cases, the expression of GCh . f / will be different from (4.1). To analyze the characteristics of the network of Figure 4.2a, however, we will refer to the study of two-terminal devices.
257
In fact, from (1.230), the cross-correlation rvi is a real function. Hence Pvi is Hermitian with even real part and odd imaginary part. Denition 4.1 The function p. f / D Re[Pvi . f /] (W/Hz) (4.5) is called average power density transferred to the load and expresses the average power per unit of frequency. We now obtain Pvi in terms of Pvb using the method shown on page 49. Being I D Vb 1 Zb C Zc V D Vb Zc Zb C Zc (4.6)
v. / i. / d
(4.2)
258
then
V I D Vb Vb
Zc jZ b C Z c j2
(4.7)
In general, if v is the voltage at the load impedance Z c , the following relation holds p. f / D Pv . f / Rc jZ c j2 (4.10)
Denition 4.2 The available power per unit of frequency of an active two-terminal device is dened as the maximum of (4.9) with respect to Z c and is obtained for Z c D Z b : pd . f / D Pvb . f / 4Rb Rb D Re[Z b ] (4.11)
We note that (4.11) is a parameter of the active device that expresses the maximum power per unit of frequency that can be delivered to the load. Active two-terminal device as a current generator. For the circuit of Figure 4.4, that consists of a current source with admittance Yb D G b C j Bb and a load with Yc D Yb , the available power per unit of frequency is given by pd . f / D Pib . f / 4G b (4.12)
where G b D Re[Yb ], and Pib (A2 /Hz) is the PSD of the signal i b .
259
We have a simple relation between the circuit of Figure 4.4 and that of Figure 4.3 for Z b D Rb C j X b ; from Norton theorem we get Yb D and Ib D Vb 1 Zb (4.14) 1 Rb D Zb jZ b j2 j Xb jZ b j2 (4.13)
where B is the passband of vb (see Denitions 1.10 on page 29 and 1.11 on page 46). According to Heaviside conditions (1.144), the voltage vb is transferred to the load without distortion if Zb. f / D K1 Zc. f / f 2B (4.16)
where K 1 is a constant. In a connection between a source and a load, the conditions for the absence of signal distortion, as well as for maximum transfer of power to the load, are veried in the following two cases. 1. For a broadband signal vb , regarding the impedances as complex-valued functions of the frequency within the passband B, the only way to verify the conditions Z c . f / D Z b . f / and (4.16) is that both the load impedance (Rc ) and the source impedance (Rb ) are purely resistive. Note that for Rb D Rc also the condition for maximum transfer of power is satised. 2. For a narrowband signal vb , for which the impedances are regarded as complexvalued constants within the passband B, the condition for maximum transfer of power Z c D Z b is easily veried. Distortion is not a problem because G. f / is a complex constant in the passband of vb , and the phase term is equivalent to a delay.
260
and G1
o.
f/ D
Vo . f / V1 . f /
(4.19)
The conditions for the absence of distortion between vi and v L are veried if G. f / D VL D Gi . f /G1 Vi
o.
f /G L . f /
(4.20)
is a constant in the passband of vi . Note that also in this case the presence of a constant delay factor is possible. Let pi . f / and po . f / be the average power densities of source and load, respectively: pi . f / D Pv1 . f / and po . f / D Pv L . f / RL RL D Pvo . f / jZ L j2 jZ L C Z 2 j2 (4.22) R1 R1 D Pvi . f / 2 jZ 1 j jZ 1 C Z i j2 (4.21)
Denition 4.3 The network power gain is dened as the ratio g. f / D po . f / pi . f / (4.23)
Using the expressions of the various frequency responses, and observing (see Figure 4.2c) Pv L . f / D jGCh . f /j2 Pv1 . f / we get g. f / D Pv L . f / 1 jZ 1 j2 RL R L jZ 1 j2 D jGCh . f /j2 R1 jZ L j2 jZ L j2 Pv1 . f / R1 (4.25) (4.24)
In the presence of match for maximum transfer of power only at the source, we introduce the notion of transducer gain, dened as gt . f / D po . f / pi;d . f / (4.26)
where pi;d . f / is the available power per unit of frequency at the source. Denition 4.4 A 2-port network is said to be perfectly matched if the conditions for maximum transfer of power are established at the source as well as at the load: Z 1 D Z i and
Z L D Z2
(4.27)
261
In this case the powers become available powers: at the source at the load pi;d . f / D Pvi . f / po;d . f / D Pvo . f / 1 4Ri 1 4R2 (4.28) (4.29)
Denition 4.5 The available power gain is dened as the ratio between po;d . f / and pi;d . f /, gd . f / D po;d . f / pi;d . f / (4.30)
In particular for Z 1 D Z 2 , that is when input and output network impedances coincide, from (4.25) we get gd . f / D jGCh . f /j2 (4.31) If in the passband B of vi we have gd > 1 the network is said to be active. If instead gd < 1 then the network is passive; in this case, we speak of available attenuation of the network : 1 (4.32) ad D gd In dB, .gd /dB D 10 log10 gd and .ad /dB D .gd /dB (4.34) (4.33)
Denition 4.6 Apart from a possible delay, for an ideal distortionless network with power gain (attenuation) gd .ad /, we will assume that the frequency response of the network is GCh . f / D Go constant f 2B (4.35)
Consequently, the impulse response is given by gCh .t/ D Go .t/ where, observing (4.31), Go D 1 p gd D p ad (4.37) (4.36)
We note that, in case the conditions leading to (4.31) are not veried, the relation between Go and gd is more complicated (see (4.25)).
262
W),
With reference to (4.37), we note that .Go /d B D .gd /d B . In fact, as Go denotes a ratio of voltages, it follows that .Go /d B D 20 log10 Go D 10 log10 gd D .gd /d B (4.42)
For telephone signals, a further power unit is given by dBrnc, which expresses the power in dBrn of a signal ltered according to the mask given in Figure 4.5 [2]. The lter reects the perception of the human ear and is known as C-message weighting.
Figure 4.5. Frequency weighting known as C-message weighting. [ c 1982 Bell Telephone Laboratories. Reproduced with permission of Lucent Technologies, Inc./Bell Labs.]
263
4.2
Various noise and disturbance signals are added to the desired signal at different points of a transmission system. In addition to interference caused by electromagnetic coupling between various system elements and noise coming from the surrounding environment, there is also noise generated by the transmission devices themselves. Such noise is very important because it determines the limits of the system. We will analyze two types of noise generated by transmission devices: thermal noise and shot noise.
Thermal noise
Thermal noise is a phenomenon associated with Brownian or random motion of electrons in a conductor. As each electron carries a unit charge, its motion between collisions with atoms produces a short impulse of current. Actually, if we represent the motion of an electron within a conductor in a two-dimensional plane, the typical behavior is represented in Figure 4.6a where the changes in the direction of the electron motion are determined by random collisions with atoms at the set of instants ftk g. Between two consecutive collisions the electron produces a current that is proportional to the projection of the velocity onto the axis of the conductor. For example, the behavior of instantaneous current for the path of Figure 4.6a is illustrated in Figure 4.6b. Although the average value (DC component) is zero, the large number of electrons and collisions gives origin to a measurable alternating component. If a current ows through the conductor, an orderly motion is superimposed on the disorderly motion of electrons; the sources of the two motions do not interact with each other. For a conductor of resistance R, at an absolute temperature of T Kelvin, the power spectral density of the open circuit voltage w at the conductor terminals is given by Pw . f / D 2kTR . f / where k D 1:3805 10
23
(4.43)
(4.44)
Figure 4.6. Representation of electron motion and current produced by the motion.
264
where h D 6:6262 10 34 Js is the Planck constant. We note that, for f kT= h D 6 1012 Hz (at room temperature T D 290 K), we get . f / ' 1. Therefore the PSD of w is approximately white, i.e. Pw . f / D 2kTR (4.45)
We adopt the electrical model of Figure 4.7, where a conductor is modelled as a noiseless device having in series a generator of noise voltage w.2 Because at each instant the noise voltage w.t/ is due to the superposition of several current pulses, a suitable model for the amplitude distribution of w.t/ is the Gaussian distribution with zero mean. Note that the variance is very large, because of the wide support of Pw . In the case of a linear two-terminal device with impedance Z D R C j X at absolute temperature T, the spectral density of the open circuit voltage w is still given by (4.43), where R D Re[Z ]. In other words, only the resistive component of the impedance gives origin to thermal noise. Let us consider the scheme of Figure 4.8, where a noisy impedance Z D R C j X is matched to the load for maximum transfer of power. Observing (4.11), the available noise power per
Figure 4.8. Electrical circuit to measure the available source noise power to the load.
An equivalent model assumes a noiseless conductor in parallel to a generator of noise current j .t/ with PSD 1 P j . f / D 2kT R . f /.
265
(4.46)
177 (dBm/Hz)
If the circuit of Figure 4.8 has a bandwidth B, the power delivered to the load is equal to Pw D kT 2B D kTB (W) 2 (4.48)
We note that a noisy impedance produces an open circuit voltage w with a root mean-square (rms) value equal to p p p w D Pw 2B D pw;d 4R2B D kT 4R B (V) (4.49) We also note from (4.48) that the total available power of a thermal noise source is proportional to the product of the system bandwidth and the absolute temperature of the source. For T D 290 K, .Pw /dBm D 174 C 10 log10 B (dBm) (4.50)
Shot noise
Most devices are affected by shot noise, which is due to the discrete nature of electron ow: also in this case the noise represents the instantaneous random deviation of current or voltage from the average value. Shot noise, expressed as a current signal, can also be modelled as Gaussian noise with a constant PSD given by Pishot . f / D eI (A2 /Hz) (4.51)
where e D 1:6 10 19 C is the electron charge and I is the average current that ows through the device; in this case it is convenient to use the electrical model of Figure 4.4.
266
In other words, Tw represents the absolute temperature that a thermal noise source should have in order to produce the same available noise power as the device. This concept can be extended and applied to the output of an amplier or an antenna, expressing the noise power in terms of effective noise temperature. We note that if a device at absolute temperature T contains more than one noise source, then Tw > T.
Figure 4.9. Noise source connected to a noisy 2-port network: three equivalent models.
267
signal given by wo D wo.S/ C wo.A/ . Assuming the two noise signals wo.S/ and wo.A/ are uncorrelated, the available power at the load will be equal to the sum of the two powers, i.e. p wo . f / D g d . f / kT S C pwo.A/ . f / 2 (4.54)
Denition 4.7 The effective noise temperature T A of the 2-port network is dened as TA. f / D pwo.A/ . f / gd . f / k 2 (4.55)
and denotes the temperature of a thermal noise source connected to a 2-port noiseless network that produces the same output noise power. Then (4.54) becomes:3 k pwo . f / D gd . f / [T S C T A ] 2 (4.56)
Denition 4.8 The effective input temperature of a system consisting of a source connected to a 2-port network is Twi D T S C T A (4.57)
Denition 4.9 The effective output temperature of a system consisting of a source connected to a 2-port network is Two D gd . f /Twi Then p wo . f / D k Tw 2 o (4.59) (4.58)
Equivalent-noise models
By the previous considerations, we introduce the equivalent circuits illustrated in Figure 4.9b and 4.9c. In particular the scheme of Figure 4.9b assumes the network to be noiseless and an equivalent noise source is considered at the input. The scheme of Figure 4.9c, on the other hand, considers all noise sources at the output. The effects on the load for the three schemes of Figure 4.9 are the same.
To simplify the notation we have omitted indicating the dependency on frequency of all noise temperatures. Note that the dependency on frequency of T A and T S is determined by gd . f /, pwo.A/ . f /, and pwi.S/ . f /.
268
F. f / D
(4.61)
D1C
Being pwo.S0 / . f / D gd . f / kT0 , and substituting for pwo.A/ the expression (4.55), we obtain 2 the important relation F. f / D 1 C TA T0 (4.62)
We note that F is always greater than 1 and it equals 1 in the ideal case of a noiseless 2-port network. Moreover, F is a parameter of the network and does not depend on the noise temperature of the source to which it is connected. From (4.61) the noise power of the 2-port network can be expressed as pwo.A/ . f / D k .F 2 1/T0 gd (4.63)
From the above considerations we deduce that to describe the noise of an active 2-port network, we must assign the gain gd and the noise gure F (or equivalently the noise temperature T A ). We now see that for a passive network at temperature T0 , it is sufcient to assign only one of the two parameters. Let us consider a passive network at temperature T0 , as for example a transmission line, for which gd < 1. To determine the noise gure let us assume as source an impedance, which is matched to the network for maximum transfer of power, at temperature T0 . Applying Thevenin theorem to the network output, the system is equivalent to a two-terminal device with impedance Z 2 at temperature T0 .
Given an electrical circuit, a useful relation to determine F, equivalent to (4.61), that employs the PSDs of the output noise signals, is given by (see (4.9)) F. f / D Pwo.A/ . f / Pwo . f / D1C Pw 0 . f / Pw 0 . f /
o.S / o.S /
(4.60)
269
Assuming the load is matched for maximum transfer of power, i.e. Z 2 D Z , from (4.46) L at the output we have pw0 . f / D .kT0 =2/. On the other hand, pwi.S0 / . f / D .kT0 =2/, and pw0.S0 / . f / D gd pwi.S0 / . Hence from the rst of (4.61) we have F. f / D 1 D ad gd (4.64)
where ad is the power attenuation of the network. Note that also in this case, given F, we can determine the effective noise temperature of the network, T A , according to (4.62). Summarizing, in a connection between a source and a 2-port network, the effective input temperature of the system can be expressed as Twi D T S C T A D T S C .F and the available noise power at the load is given by p w0 . f / D g d . f / kTwi 2 (4.66) 1/T0 (4.65)
Example 4.2.1 Let us consider the conguration of Figure 4.10a, where an antenna with noise temperature T S is connected to a pre-amplier with available power gain g and noise gure F. An electrical model of the connection is given in Figure 4.10b, where the antenna is modelled as a resistance with noise temperature T S . If the impedances of the two devices are matched, Twi is given by (4.65).
(a)
(b)
270
With regard to the noise characteristics, it is sufcient to determine the noise gure of the cascade of the two networks. For a source at room temperature T0 , from (4.66) for T S D T0 , the noise power at the output of the rst network is given by pwo;1 . f / D kT0 F1 g1 2 (4.68)
At the output of the second network we have pwo;2 . f / D pwo;1 . f /g2 C k .F2 2 1/T0 g2 (4.69)
using (4.63) to express the noise power due to the second network only. Then the noise gure of the overall network is given by pwo;2 . f / FD D pwo;2.S0 / . f / Simplifying (4.70) we get F D F1 C .F2 1/ g1 (4.71) kT0 kT0 g1 g2 F1 C .F2 2 2 kT0 g1 g2 2 1/g2 (4.70)
Extending this result to the cascade of N 2-port networks, Ai , i D 1; : : : ; N , characterized by gains gi and noise gures Fi , we obtain Friis formula of the total noise gure F D F1 C .F2 1/ .F3 1/ .F N 1/ C C C g1 g1 g2 g1 g2 : : : g N (4.72)
1
271
We observe that F strongly depends on the gain and noise gure parameters of the rst stages; in particular, the smaller F1 and the larger g1 , the more F will be reduced. Substituting (4.62), that relates the noise gure to the effective noise temperature, in (4.72), we have that the equivalent noise temperature of the cascade of N 2-port networks, characterized by noise temperatures T Ai , i D 1; : : : ; N , is given by T A D T A1 C T A2 T A3 T AN C C C g1 g1 g2 g1 g2 : : : g N (4.73)
1
Example 4.2.2 The idealized conguration of a transmission medium consisting of a very long cable where ampliers are inserted at equally spaced points is illustrated in Figure 4.12. Each section of the cable, with power attenuation ac and noise gure Fc D ac (see (4.64)), cascaded with an amplier, with gain g A and noise gure F A , is called a repeater section. To compensate for the attenuation of the cable we choose g A D ac . Then, each section has a gain gsr D and noise gure Fsr D Fc C FA 1 D ac C ac .F A gc 1/ D g A F A (4.76) 1 gA D 1 ac (4.75)
Therefore the N sections have overall unit gain and noise gure F D Fsr C .Fsr 1/ .Fsr 1/ C C gsr gsr gsr : : : gsr D N .Fsr 1/ C 1 ' N Fsr
(4.77)
where Fsr is given by (4.76). We note that the output noise power of N repeater sections is N times the noise power introduced by an individual section.
272
4.3
On the other hand, the effects of the two signals on a certain load Z c are measured by the average powers. Therefore we also introduce the following signal-to-noise ratio of average powers Z C1 ps . f / d f Ps 1 D Z C1 3p D (4.80) Pw pw . f / d f
1
pw . f / D Pw . f /
Therefore 3s and 3 p are in general different. However, if the term Rc =jZ b C Z c j2 is a constant within the passband of s and w, then the two SNRs coincide. Note that if Z b D Z c , 2 D that is the condition for maximum transfer of power is satised, then Rc =jZ b C Z c j 1=.4Rb /. Hence it is sufcient that Rb is constant within the passband of s and w to have 3 D 3s D 3 p (4.83) Moreover, assuming pw is constant within the passband of w, with bandwidth B, we have k Pw D Tw 2B (4.84) 2 and, from (4.80), 3D Ps E[s 2 .t/] D 2 .t/] kTw B E[w (4.85)
where Ps is the available average power of the desired signal, and Tw is the noise temperature of w. Later we will often use this relation.
273
where so and wo depend on s and wi , respectively. Under matched load conditions (that is Z L D Z 2 ), and assuming (4.83) holds, at the network output we obtain 3out D
2 Ps E[so .t/] D o 2 .t/] P wo E[wo
(4.88)
We indicate with B the passband of the network frequency response, usually equal to or including the passband of s, and with B its bandwidth. From the expressions (4.4) and (4.30) Z C1 Z Pso D pso . f / d f D 2 ps . f /g. f / d f (4.89)
1
and Z P wo D 2
B
pwi . f /g. f / d f
(4.90)
Assuming now that g. f / is constant within B, we have k Tw g2B (4.91) 2 i assuming that also the source is matched for maximum transfer of power. Finally we get Pso D Ps g and Pwo D 3out D
2 Ps E[so .t/] D 2 .t/] kTwi B E[wo
(4.92)
where Ps is the available power of the desired signal at the network input, and Twi D T S C .F 1/T0 is the effective noise temperature including both the source and the 2-port network. With reference to the above conguration, we observe that the power of wi could be very high if Twi is constant over a wide band, but wo has much smaller power since its passband coincides with that of the network frequency response. From (4.91) and (4.50), the effective input noise due to the connection source-network has an average power for T S D T0 .Twi D FT0 / equal to .Pwi /dBm D 114 C 10 log10 B jMHz C.F/d B .T S D T0 / (4.93)
and the average power of the effective output noise is given by .Pwo /dBm D .Pwi /dBm C .g/d B .T S D T0 / In (4.93), BjMHz denotes the bandwidth in MHz. (4.94)
274
Example 4.3.1 A station, receiving signals from a satellite, has an antenna with gain gant of 40 dB and a noise temperature T S of 60 K (that is the antenna acts as a noisy resistor at a temperature of 60 K). The antenna feeds a preamplier with a noise temperature T A1 of 125 K and a gain g1 of 20 dB. The preamplier is followed by an amplier with a noise gure F2 of 10 dB and a gain g2 of 80 dB. The transmitted signal bandwidth is 1 MHz. The satellite has an antenna with a power gain of gsat D 6 dB and the total attenuation a` due to the distance between transmitter and receiver is 190 dB. We want to nd: 1. the average power of the thermal noise at the receiver output, 2. the minimum power of the signal transmitted by the satellite to obtain a SNR of 20 dB at the receiver output. The two receiver ampliers can be modelled as one amplier with gain: .g A /d B D .g1 /d B C .g2 /d B D 20 C 80 D 100 dB and effective noise temperature: T A D T A1 C T A2 .F2 1/T0 .1010=10 1/290 D T A1 C D 125 C D 151 K g1 g1 1020=10 (4.96) (4.95)
1. From (4.91) the average power of the output noise is Pwo D k.T S C T A /g A B D 1:38 10 D 2:91 10
23 5
WD
(4.97)
2. From 3out D .Pso =Pwo / 20 dB we get .Pso =Pwo / 100. As Pso D Ps gsat .1=a` / gant g A D Ps 10 44=10 , it follows Ps 73 W (4.98)
A more useful relation is obtained assuming that g. f / is a constant within the passband B of the network. Given the average power of the noise generated by the source at room temperature Pwi.S0 / D kT0 2B 2 (4.100)
275
g (dB) 20 30 20 30 50
and 3in D then, from (4.92), we have 3in (4.102) .T S D T0 / F In other words, F is a measure of the reduction of the SNR at the output due to the noise introduced by the network. In Table 4.1 the typical values of F, T A , and gain g are given for three devices. In the last column the frequency range usually considered for the operations of each device is also given. 3out D Ps Pwi.S0 / (4.101)
4.4
4.4.1
Transmission lines
Fundamentals of transmission line theory
In this section, the principles of signal propagation in transmission lines are briey reviewed. A uniform transmission line consists of a two-conductor cable with a uniform cross-section, that supports the propagation of transverse electromagnetic (TEM) waves [3, 1]. Examples of transmission lines are twisted-pair cables and coaxial cables. We now develop the basic transmission line theory. With reference to Figure 4.13, which illustrates a uniform line, let
276
rdx
ldx
i+
i dx x
v gdx cdx
v+
v dx x
x denote the distance from the origin and L be the length of the line. The termination is found at distance x D 0 and the signal source at x D L. Let v D v.x; t/ and i D i.x; t/ be, respectively, the voltage and current at distance x at time t. To determine the law that establishes the voltage and current along the line, let us consider a uniform line segment of innitesimal length that we assume to be time invariant, depicted in Figure 4.14. The parameters r; `; g; c are known as primary constants of the line. They dene, respectively, resistance, inductance, conductance and capacitance of the line per unit length. Primary constants are in general slowly time-varying functions of the frequency; however, in this context, they will be considered time invariant. The model of Figure 4.14 is obtained using the rst order Taylor series expansion of v.x; t/ and i.x; t/ as a function of distance x.
(4.103)
(4.104)
Substituting @ 2 i=@t@ x in the rst equation with the expression obtained from the second, we get the wave equation @ 2v 1 @ 2v @ 2v D `c 2 D 2 2 @x2 @t @t (4.105)
277
p where D 1= `c represents the velocity of propagation of the signal on a lossless transmission line. The general solution to the wave equation for a lossless transmission line is given by x x C '2 t C (4.106) v.x; t/ D '1 t where '1 and '2 are arbitrary functions. Noting that from (4.103) @i =@t D (4.106) yields ` @i D @t 1 0 ' t 1 x 1 0 x C '2 t C .1=`/@v=@ x,
(4.107)
0 0 where '1 and '2 are the derivatives of '1 and '2 , respectively. Integrating by parts (4.107) we get
i.x; t/ D
1 h '1 t `
'2 t C
x i C '.x/
(4.108)
where '.x/ is time independent and can therefore be ignored in the study of propagation. Dening the characteristic impedance of a lossless transmission line as r ` Z 0 D ` D (4.109) c the expression for the current is given by i.x; t/ D 1 h '1 t Z0 x '2 t C x i (4.110)
From the general solution to the wave equation we nd that the voltage (or the current), considered as a function of distance along the line, consists of two waves that propagate in opposite directions: the wave that propagates from the source to the line termination is called the source or incident wave, that which propagates in the opposite direction is called reected wave. We consider now the propagation of a sinusoidal wave with frequency f D !=2 in an ideal transmission line. The voltage at distance x D 0 is given by v.0; t/ D V0 cos.!t/ (4.111)
The wave propagating in the positive direction of x is given by vC .x; t/ D jVC j cos[!.t x=/], that propagating in the negative direction is given by v .x; t/ D jV j cos[!.t C x=/ C p ]. The transmission line voltage is obtained as the sum of the two components and is given by i h h x i x (4.112) C jV j cos ! t C C p v.x; t/ D jVC j cos ! t The current has the expression h jVC j i.x; t/ D cos ! t Z0 x i i h jV j x C p cos ! t C Z0
(4.113)
278
Let us consider a point on the x-axis individuated at each time instant t by the condition that the argument of the function F.t x=/ is a constant. This point is seen by an observer as moving at velocity in the positive direction of the x-axis. For sinuosoidal waves the velocity for which the phase is a constant is called phase velocity . It is useful to write (4.112) and (4.113) in complex notation, where the phasors V and I represent amplitude and phase at distance x of the sinusoidal signals (4.112) and (4.113), respectively, V D VC e jx C V e jx 1 ID .VC e jx V e jx / Z0 (4.114) (4.115)
where D != denotes the phase constant. We dene the wavelength as D 2=. We note that frequency f and wavelength are related by D f (4.116)
In particular, the propagation in free space is characterized by D c D 3 108 m/s. If VC is taken as the reference phasor with phase equal to zero, then V D jV je j p , where p is the phase rotation between the incident and the reected waves at x D 0. Let us consider a transmission line having as termination an impedance Z L . By Kirchhoff laws, the voltage and current at the termination are given by 8 > VL D VC C V < (4.117) V > I L D VL D VC : ZL Z0 Z0 The reection coefcient is dened as the ratio between the phasors representing, respectively, the reected and incident waves, % D V =VC . The transmission coefcient is dened as the ratio between the phasors representing, respectively, the termination voltage and the incident wave D VL =VC . From (4.117), it turns out V Z L Z0 %D D e j p (4.118) V Z L C Z0 C and D 2Z L Z L C Z0 (4.119)
At the termination, dening the incident power as PC D jVC j2 =Z 0 and the reected power as P D jV j2 =Z 0 , we obtain P =PC D j%j2 ; the ratio between the power delivered to the load and the incident power is hence given by 1 j%j2 . Let us consider some specic cases: if Z L D Z 0 , % D 0 and there is no reection; if Z L D 1, the line is open-circuited, % D 1 and V D VC ; if Z L D 0, the line is short-circuited, % D 1 and V D VC .
279
(4.120)
Differentiating and substituting in the rst equation the expression of d I =dx obtained from the second, we get d2V D dx 2 where p D ZY (4.122) is a characteristic constant of the transmission line called propagation constant. Let and be, respectively, the real and imaginary parts of : is the attenuation constant measured in neper per unit of length, and is the phase constant measured in radians per unit of length. The solution of the differential equation for the voltage can be expressed in terms of exponential functions as V D VC e
x 2
(4.121)
CV e
D VC e
jx
C V ex e jx
(4.123)
V e
(4.124)
is the characteristic impedance of the transmission line. The propagation constant and the characteristic impedance are also known as secondary constants of the transmission line.
Frequency response
Let us consider the transmission line of Figure 4.15, with a sinusoidal voltage source vi and a load Z L . From (4.123) the voltage at the load can be expressed as VL D VC .1 C %/. Recalling that V = VC D %, we dene the voltage Vo D VL j Z LD1 D VC .1 C %/ j%D1 D 2VC .
280
i(t)
1
v 1 (t) v(t)
i
v L(t)
ZL
x = -L
x=0
Figure 4.15. Transmission line with sinusoidal voltage generator vi and load ZL .
C %e
L/
(4.126)
where Z i denotes the generator impedance. The input and output impedances of the 2-port network are, respectively, given by: Z1 D Z2 D 1 C %e V1 D Z0 I1 1 %e
2 L 2 L
(4.127) (4.128)
where I L sc D I L j Z L D0 and Vooc D VL j Z L D1 . We now want to determine the ratio between the voltage VL and the voltage V1 , dened as GCh D VL = V1 . Observing the above relations we nd the following frequency responses: GL D G1 D ZL 1 VL 1C% D D VC .1 C %/ D Vo 2VC 2 Z L C Z0 Vo 2e D V1 1 C %e
L 2 L
(4.129)
(4.130)
Gi D
V1 Z1 Z 0 .1 C %e 2 L / D D Vi Zi C Z1 Z 0 .1 C %e 2 L / C Z i .1 %e
2 L/
(4.131)
281
Then, from (4.1), the channel frequency response is given by: GCh D G1 Let us consider some specic cases: Matched transmission line: % D 0 for Z i D Z L D Z 0 .Gi D 1=2/ GCh D e Short-circuited transmission line: % D 1 (4.134)
L o GL
.1 C %/e 1 C %e 2
L L
(4.132)
(4.133)
1 cosh. L/
(4.135)
To determine the power gain of the network, we can use the general equation (4.25), or observe (4.23); in any case, we obtain g. f / D 1 j%j2 e 1 j%j2 e 4L
2L
(4.136)
where D Re[ ]. We note that, for a matched transmission line, the available attenuation is given by ad . f / D 1 je
L j2
D e2L
(4.137)
In (4.137), expresses the attenuation in neper per unit of length. Alternatively, one can introduce an attenuation in dB per unit of length, .ad . f //d B , as Q
Q ad . f / D 10 10 .ad . f //d B L 1
(4.138)
The relation between and .ad . f //d B is given by Q .ad . f //d B D 8:68 Q (4.139)
From (4.139), the attentuation in dB introduced by the transmission line is equal to .ad . f //d B D .ad . f //d B L Q (4.140)
In a transmission line with a non-matched resistive load that satises the condition Z L Z 0 , from (4.118) we get 1 C % ' 2Z L =Z 0 , %2 e 4L ' 0, and %2 ' 1 4Z L =Z 0 . Therefore (4.136) yields ZL .ad . f //d B D .ad . f //d B L 10 log10 4 (4.141) Q Z 0
282
For frequencies at which r !`, using the approximation 1=2 r2 1 r2 1C 2 2 '1C 2 ! 2 `2 ! ` we nd r ' 2 r c ` p and ' ! `c (4.146) (4.145)
283
and the attenuation introduced by the transmission line can be expressed as p .ad . f //d B D 8:68K f (dB/m) Q (4.149)
We note that, given the value of . f / at a certain frequency f D f 0 , we can obtain the value of K . Therefore it is possible to determine the attenuation constant at every other frequency. From the expression (4.133) of the frequency response of a matched transmission p line, with given by (4.147), without considering the delay `c introduced by the term p j! `c, the impulse response has the following expression KL gCh .t/ D p e 2 t3
.K L/2 4t
1.t/
(4.150)
The pulse signal gCh is shown in Figure 4.16 for various values of the product K L. We note a larger dispersion of gCh for increasing values of K L.
KL=2
0.1
0.08
gCh(t)
0.06
KL=3
0.04
KL=4
0.02
KL=5 KL=6
8 t (s)
10
12
14
16
Figure 4.16. Impulse response of a matched transmission line for various values of KL.
284
Frequency (Hz) 1000 2000 3000 1000 2000 3000 1000 2000 3000 1000 2000 3000
Characteristic impedance Z0 ( ) 297 217 183 414 297 247 518 370 306 654 466 383 j278 j190 j150 j401 j279 j224 j507 j355 j286 j645 j453 j367
Propagation constant C j (neper/km) (rad/km) 0:09 C 0:12 C 0:15 C 0:13 C 0:18 C 0:22 C 0:16 C 0:23 C 0:28 C 0:21 C 0:29 C 0:35 C j0:09 j0:14 j0:18 j0:14 j0:19 j0:24 j0:17 j0:24 j0:30 j0:21 j0:30 j0:37
Attenuation ad D 8:68 Q (dB/km) 0:78 1:07 1:27 1:13 1:57 1:90 1:43 2:00 2:42 1:81 2:55 3:10
c 1982 Telephone Laboratories. Reproduced with permission of Lucent Technologies, Inc./Bell Labs.
Figure 4.17. Attenuation as a function of frequency for some telephone transmission lines: three are polyethylene-insulated cables (PIC) and one is a coaxial cable with a diameter of 9.525 mm. [ c 1982 Bell Telephone Laboratories. Reproduced with permission of Lucent Technologies, Inc./Bell Labs.]
285
albeit with a different constant of proportionality. In any case in the local-loop,5 to force the primary constants to satisfy Heaviside conditions in the voice band, which goes from 300 to 3400 Hz, formerly some lump inductors were placed at equidistant points along the transmission line. This procedure, called inductive loading, causes . f / to be at in the voice band, but considerably increases the attenuation outside of the voice band. Moreover, the phase . f / may result very distorted in the passband. Typical behavior of and in the frequency band 0 4000 Hz, with and without loading, are given in Figure 4.18 for a transmission line with gauge 22 [2]. The digital subscriber line (DSL) technologies, introduced for data transmission in the local loop, require a bandwidth much greater than 4 kHz, up to about 20 MHz for the VDSL technology (see Chapter 17). For DSL applications it is therefore necessary to remove possible loading coils that are present in the local loops. The frequency response of a DSL transmission line can also be modied by the presence of one or more bridged-taps. A bridged-tap consists of a twisted pair cable of a certain length L BT , terminated by an open circuit and connected in parallel to a local loop. At the connection point, the incident signal separates into two components. The component propagating along the bridged-tap is reected at the point of the open circuit: the component propagating on the transmission line must therefore be calculated taking also into consideration this reected component. At the frequencies f BT D = BT , where BT satises the condition .2n C 1/ BT =4 D L BT , n D 0; 1; : : : , at the connection point we get destructive interference between the reected and incident component: this interference reveals itself as a notch in the frequency response of the transmission line. Given
Figure 4.18. Attenuation constant and phase constant for a telephone transmission line with and without loading. [ c 1982 Bell Telephone Laboratories. Reproduced with permission of Lucent Technologies, Inc./Bell Labs.]
By local-loop we intend the transmission line that goes from the user telephone set to the central ofce.
286
Table 4.3 Transmission characteristics dened by the EIA/TIA for unshielded twisted pair (UTP) cables.
Signal attenuation at 16 MHz UTP-3 UTP-4 UTP-5 13.15 dB/100 m 8.85 dB/100 m 8.20 dB/100 m
the large number of transmission lines actually in use, to evaluate the performance of DSL systems we usually refer to a limited number of loop characteristics, which can be viewed as samples taken from the ensemble of frequency responses. On the other hand, the transmission characteristics of unshielded twisted-pair (UTP) cables commonly used for data transmission over local area networks are dened by the EIA/TIA and ISO/IEC standards. As illustrated in Table 4.3, the cables are divided into different categories according to the values of 1) the signal attenuation per unit of length, 2) the attenuation of the near-end cross-talk signal, or NEXT, that will be dened in the next section, and 3) the characteristic impedance. Cables of category three (UTP-3) are commonly called voice-grade, those of categories four and ve (UTP-4 and UTP-5) are data-grade. We note that the signal attenuation and the intensity of NEXT are substantially larger for UTP-3 cables than for UTP-4 and UTP-5 cables.
4.4.2
Cross-talk
The interference signal that is commonly referred to as cross-talk is determined by magnetic coupling and unbalanced capacitance between two adjacent transmission lines. Let us consider the two transmission lines of Figure 4.19, where the terminals .1; 10 / belong to the disturbing transmission line and the terminals .2; 20 / belong to the disturbed transmission line. In the study of the interference signal produced by magnetic coupling, we consider
287
i1 1 v1 Z0 1 im 2 Z0 2
Figure 4.20. Interference signal produced by magnetic coupling.
1 c 11 v1 c 12 1 2 Z0 2 1 (a) (b) c1 2 c1 2 c 22 Z0 c1 2 Z0 c1 2 c 12 Z0 v1 Z 0 c 11 2 c 12 ic 1 Z0 c 22 c 12 2
Z0
the circuit of Figure 4.20. We will assume that the length of the transmission line is much longer than the wavelength corresponding to the maximum transmitted frequency and that the impedance Z 0 is much higher than the inductor reactance. The induced electromagnetic force (EMF) is given by E D j2 f m I1 , where I1 ' V1 =Z 0 . The EMF produces a current j2 f E Im D .1=.2Z 0 // D .1=.2Zm// I1 , that can be expressed as Im D j2 f m V1 . 2 0 To study the interference signal due to unbalanced capacitance, we consider the circuit of Figure 4.21a, that can be redrawn in an equivalent way as illustrated in Figure 4.21b. We assume that the impedance Z 0 is much smaller than the reactance of the capacitors that can be found on the bridge. Applying the principle of the equivalent generator we nd 0 Ic D V220 j Ic D0 D Z 220 1 1 c10 2 1 B c10 20 B @ 1 1 C c10 20 c120 C 1 C j2 f V1 A 1 1 1 1 C C c10 2 c12 c12 C c10 2 c120 C c10 20 (4.151)
.1=.2Z 0 //
288
Figure 4.22. Illustration of near-end cross-talk (NEXT) and far-end cross-talk (FEXT) signals.
from which we obtain Ic D c12 c10 20 c120 c10 2 j2 f V1 D j2 1cV1 c12 C c10 2 C c120 C c10 20 (4.152)
Recalling that the current Ic is equally divided between the impedances Z 0 on which the transmission line terminates, we nd that the cross-talk current produced at the transmitter side termination is I p D Im C Ic =2, and the cross-talk current produced at the receiver side termination is It D Im C Ic =2. As illustrated in Figure 4.22, the interference signals are called near-end cross-talk or NEXT, or far-end cross-talk or FEXT, depending on whether the receiver side of the disturbed line is the same as the transmitter side of the disturbing line, or the opposite side, respectively. We now evaluate the total contribution of the near and far-end cross-talk signals for lines with distributed impedances.
Near-end cross-talk
Let a p .x/ D m.x/ 1c.x/ Z0 C 2Z 0 2 (4.153)
be the near-end cross-talk coupling function at distance x from the origin. In complex notation, the NEXT signal is expressed as Z Vp D Z0 I p D
0 L
V1 e
2 x
j2 f a p .x/ dx
(4.154)
To calculate the power spectral density of NEXT we need to know the autocorrelation function of the random process a p .x/. A model commonly used in practice assumes that a p .x/ is a white stationary random process, with autocorrelation ra p .z/ D E[a p .x C z/a .x/] D r p .0/.z/ p (4.155)
289
For NEXT the following relation holds E[jV p . f /j2 ] D E[jV1 . f /j2 ] 3=2 r p .0/ f 3=2 .1 K e4K
p f L
Using (1.449), the level of NEXT coupling is given by6 jG p . f /j2 D E[jV p . f /j2 ] ' k p f 3=2 E[jV1 . f /j2 ] (4.158)
To perform computer simulations of data transmission systems over metallic lines in the presence of NEXT, it is required to characterize not only the amplitude, but also the phase of NEXT coupling. In addition to experimental models obtained through laboratory measurements, the following stochastic model is used:
L 1x 1 X i D0
a p .x/ D with
ai w1x .x
i1x/
(4.159)
( w1x .x/ D
1 0
(4.160)
where ai , i D 0; : : : ; L=1x 1, denote statistically independent Gaussian random variables with zero mean and variance E[ai2 ] D A NEXT coupling function is thus given by
L 1x 1 X i D0 p p p 1 2.K f C j K f C j2 f `c/.i C 2 /1x
r p .0/ 1x
(4.161)
GNE X T . f / D
j2 f ai w1x .x
i1x/e
(4.162)
If we know the parameters of the transmission line K and k p , then from (4.157) and (4.161) the variance of ai to be used in the simulations is given by E[ai2 ] D K kp 3=2 1x (4.163)
Observing (1.449), jG p . f /j2 is also equal to the ratio between the PSDs of v p and v1 .
290
Far-end cross-talk
Let at .x/ D m.x/ 1c.x/ Z0 C 2Z 0 2 (4.164)
be the far-end cross-talk coupling function at distance x from the origin. In complex notation, the FEXT signal is given by Z L V1 e L j2 f at .x/ dx (4.165) Vt D Z 0 It D
0
Analogously to the case of NEXT, we assume that at is a white stationary random process, with autocorrelation rat .z/ D E[a t .x C z/a t .x/] D rt .0/.z/ For the FEXT signal the following relation holds E[jVt . f /j2 ] D E[jV1 . f /j2 ]e
p 2K f L
(4.166)
.2 f /2 rt .0/L
(4.167)
where L is the length of the transmission line. The level of FEXT coupling is given by jGt . f /j2 D E[jVt . f /j2 ] D kt f 2 Le E[jV1 . f /j2 ]
p 2K f L
(4.168)
where kt D .2 /2 rt .0/. We note that for high-speed data transmission systems over unshielded twisted-pair cables, NEXT usually represents the dominant source of interference. Example 4.4.1 For local-area network (LAN) applications, the maximum length of cables connecting stations is typically limited to 100 m. Deviations from the characteristic expressed by (4.147) may be caused by losses in the dielectric material of the cable, the presence of connectors, non-homogeneity of the transmission line, etc. For the IEEE Standard 100BASE-T2, which denes the physical layer for data transmission at 100 Mb/s over UTP-3 cables in Ethernet LANs (see Chapter 17), the following worst-case frequency response is considered: GCh . f / D 10
p 1:2 20 e .0:00385 j f C0:00028 f /L p
(4.169)
where f is expressed in MHz and L in meters. In (4.169), the term e j2 f `cL is ignored, as it indicates a constant propagation delay. A frequency independent attenuation of 1.2 dB has been included to take into account the attenuation caused by the possible presence of connectors. The amplitude of the frequency response obtained for a cable length L D 100 m is shown in Figure 4.23 [4]. We note that the signal attenuation at the frequency of 16 MHz is equal to 14.6 dB, a higher value than that indicated in Table 4.3 for UTP-3 cables.
291
0 Amplitude characteristic for 100 m cable length 10 14.6 dB 21.0 dB 16 MHz NEXT coupling envelope curve 21 + 15 log10 ( f/16 ) , f in MHz
20 (dB) 30 Amplitude 40
60
10
15
20 f (MHz)
25
30
35
40
Figure 4.23. Amplitude of the frequency response for a voice-grade twisted-pair cable with length equal to 100 m, and four realizations of NEXT coupling function. [ c 1997 IEEE.]
The level of NEXT coupling (4.158) is illustrated in Figure 4.23 as a dotted line; we note the increase as a function of frequency of 15 dB/decade, due to the factor f 3=2 . The level of NEXT coupling equal to 21 dB at the frequency of 16 MHz is larger than that given in Table 4.3 for UTP-3 cables. The amplitude characteristics of four realizations of the NEXT coupling function (4.162) are also shown in Figure 4.23.
4.5
Optical bers
Transmission systems using light pulses that propagate over thin glass bers were introduced in the 1970s and have since then undergone continuous development and experienced an increasing penetration, to the point that they now constitute a fundamental element of modern information highways. For in-depth study of optical ber properties and of optical component characteristics we refer the reader to the vast literature existing on the subject [5, 6, 7]; in this section we limit ourselves to introducing some fundamental concepts. The term optical communications is used to indicate the transmission of information by the propagation of electromagnetic elds at frequencies typically of the order of 1014 1015 Hz, that are found in the optical band and are much higher than the frequency of radio waves or microwaves; to identify a transmission band, the wavelength rather than the frequency is normally used. We recall that for electromagnetic wave propagation in free space, the relation (4.116) holds: a frequency of 3 1014 Hz corresponds therefore to a wavelength of 1 m for transmission over optical bers. The signal attenuation as a function of the wavelength exhibits the behavior shown in Figure 4.24 [8, 9]; we note that the useful interval for transmission is in the range from 800 to 1600 nm, that corresponds
292
Figure 4.24. Attenuation curve as a function of wavelength for an optical ber. [From Li (1980), see also Miya et al. (1979), c 1980 IEEE.]
to a bandwidth of 2 1014 Hz. Three regions are typically used for transmission: the rst window goes from 800 to 900 nm, the second from 1250 to 1350 nm, and the third from 1500 to 1600 nm. We immediately realize the enormous capacity of ber transmission systems: for example, a system that uses only 1% of the 2 1014 Hz bandwidth mentioned above, has an available bandwidth of 2 1012 Hz, equivalent to that needed for the transmission of 300:000 television signals, each with a bandwidth of 6 MHz. To efciently use the band in the optical spectrum, multiplexing techniques using optical devices have been developed, such as wavelength-division multiplexing (WDM) and optical frequency-division multiplexing (O-FDM). Moreover, we note that, although the propagation of electromagnetic elds in the atmosphere at these frequencies is also considered for transmission (see Section 17.2.1), the majority of optical communication systems employ as transmission medium an optical ber, which acts as a waveguide. A fundamental device in optical communications is represented by the laser, which, beginning in the 1970s, made coherent light sources available for the transmission of signals.
293
where M is the dispersion coefcient of the material, Mg is the dispersion coefcient related to the geometry of the waveguide, L denotes the length of the ber and 1 denotes the spectral width of the light source. The total dispersion .M C Mg / has values near 120, 0, and 15 ps/(nmkm) at wavelengths of 850, 1300, and 1550 nm, respectively. The bandwidth of the transmission medium is inversely proportional to the dispersion; we note that the dispersion is minimum in the second window, with values near zero around the wavelength of 1300 nm for conventional bers. Special bers are designed to compensate for the dispersion introduced by the material; because of the low attenuation and dispersion, these bers are normally used in very long distance connections. Multimode bers allow the propagation of more than one mode of the electromagnetic eld. In this case the medium introduces signal distortion caused by the fact that propagation of energy for different modes has different speeds: for this reason multimodal bers are used in applications where the transmission bandwidth and the length of the transmission line are not large. Monomode bers limit the propagation to a single mode, thus eliminating the dispersion caused by multimode propagation. Because in this case the dispersion is due only to the material and the geometry of the waveguide, monomodal bers are preferred for applications that require wide transmission bandwidth and very long transmission lines. In Table 4.4 typical values of the transmission bandwidth, normalized by the length of the optical ber, are given for different types of bers. The step-index (SI) ber is characterized by a constant value of the refraction index, whereas the graded-index (GRIN) ber has a refraction index decreasing with the distance from the ber axis. As noticed previously, the monomodal bers are characterized by larger bandwidths; to limit the number of modes
294
to one, the diameter of the monomodal ber is related to the wavelength and is normally about one order of magnitude smaller than that of multimodal bers. Semiconductor laser diodes (LD) or light-emitting diodes (LED) are used as signal light sources in most applications; these sources are usually modulated by electronic devices. The conversion from a current signal to an electromagnetic eld that propagates along the ber can be described in terms of light signal power by the relation PT x D k0 C k1 i (4.171)
where k0 and k1 are constants. The transmitted waveform can therefore be seen as a replica of the modulation signal, in this case the current signal. Laser diodes are characterized by a smaller spectral width 1 as compared to that of LEDs, and therefore lead to a lower dispersion (see (4.170)). The more widely used photodetector devices are semiconductor photodiodes, which convert the optical signal into a current signal according to the relation i D P Rc (4.172) where i is the device output current, P Rc is the power of the incident optical signal and is the photodetector response. Typical values of are of the order of 0.5 mA/mW. Signal quality is measured by the signal-to-noise ratio expressed as 3D gin .gi P Rc /2 R L 2e R L B.I D C P Rc / C 4kTw B (4.173)
where gi is the photodetector current gain, n is a parameter that indicates the photodetector excess noise, B is the receiver bandwidth, k is Boltzmann constant, e is the charge of the electron, Tw is the effective noise temperature in Kelvin, I D is the photodetector dark current, and R L is the resistance of the load that follows the photodetector. We note that in the denominator of (4.173) the rst term is due to shot noise and the second term to thermal noise.
4.6
Radio links
The term radio is used to indicate the transmission of an electromagnetic eld that propagates in free space. Some examples of radio transmission systems are: point-to-point terrestrial links [11]; mobile terrestrial communication systems [12, 13, 14, 15, 16];
295
earth-satellite links (with satellites employed as signal repeaters) [17]; deep-space communication systems (with space probes at a large distance from earth). A radio link model is illustrated in Figure 4.26, where we assume that the transmit antenna input impedance and the receive antenna output impedance are matched for maximum transfer of power.
4.6.1
Frequencies used for radio transmission are in the range from about 100 kHz to some tens of GHz. The choice of the carrier frequency depends on various factors, among which the dimensions of the transmit antenna play an important role. In fact, to achieve an efcient radiation of electromagnetic energy, one of the dimensions of the antenna must be at least equal to 1=10 of the carrier wavelength. This means that an AM radio station, with carrier frequency f 0 D 1 MHz and wavelength D c= f 0 D 300 m, where c is the speed of light in free space, requires an antenna of at least 30 m. A radio wave usually propagates as a ground wave (or surface wave), via reection and scattering in the atmosphere (or via tropospheric scattering), or as a direct wave. Recall that, if the atmosphere is non-homogeneous (in terms of temperature, pressure, humidity, : : : ), the electromagnetic propagation depends on the changes of the refraction index of the medium. In particular, this gives origin to the reection of electromagnetic waves. We speak of diffusion or scattering phenomena if molecules that are present in the atmosphere absorb part of the power of the incident wave and then re-emit it in all directions. Obstacles such as mountains, buildings, etc., give also origin to signal reection and/or diffusion. In any case, these are phenomena that permit transmission between two points that are not in line-of-sight (LOS). We will now consider the types of propagation associated with frequency bands. Very low frequency (VLF) for f 0 < 0:3 MHz. The earth and the ionosphere form a waveguide for the electromagnetic waves. At these frequencies the signals propagate around the earth.
296
Medium frequency (MF) for 0:3 < f 0 < 3 MHz. The waves propagate as ground waves up to a distance of 160 km. High frequency (HF) for 3 < f 0 < 30 MHz. The waves are reected by the ionosphere at an altitude that may vary between 50 and 400 km. Very high frequency (VHF) for 30 < f 0 < 300 MHz. For f 0 > 30 MHz, the signal propagates through the ionosphere with small attenuation. Therefore these frequencies are adopted for satellite communications. They are also employed for line-of-sight transmissions, using high towers where the antennas are positioned to cover a wide area. The limit to the coverage is set by the earth curvature. If h is the height of the tower in meters, the range p covered expressed in km is r D 1:3 h: for example, if h D 100 m, coverage is up to about r D 13 km. However, ionospheric and tropospheric scattering (at an altitude of 16 km or less) are present at frequencies in the range 3060 MHz and 40300 MHz, respectively, which cause the signal to propagate over long distances with large attenuations. Ultra high frequency (UHF) for 300 MHz < f 0 < 3 GHz. Super high frequency (SHF) for 3 < f 0 < 30 GHz. At frequencies of about 10 GHz, atmospheric conditions play an important role in signal propagation. We note the following absorption phenomena, which cause additional signal attenuation: 1. due to oxygen: for f 0 > 30 GHz, with peak attenuation at 60 GHz; 2. due to water vapor: for f 0 > 20 GHz, with peak attenuation at around 20 GHz; 3. due to rain: for f 0 > 10 GHz, assuming the diameter of the rain drops is of the order of the signal wavelength. We note that, if the antennas are not positioned high enough above the ground, the electromagnetic eld propagates not only into the free space but also through ground waves. Extremely high frequency (EHF) for f 0 > 30 GHz.
Radiation masks
A radio channel by itself does not set constraints on the frequency band that can be used for transmission. In any case, to prevent interference among radio transmissions, regulatory bodies specify power radiation masks: a typical example is given in Figure 4.27, where the plot represents the limit on the power spectrum of the transmitted signal with reference to the power of a non-modulated carrier. To comply with these limits, a lter is usually employed at the transmitter front-end.
4.6.2
The propagation of electromagnetic waves should be studied using Maxwell equations with appropriate boundary conditions. Nevertheless, for our purposes a very simple model, which
297
Figure 4.27. Radiation mask of the GSM system with a bandwidth of 200 kHz around the carrier.
consists in approximating an electromagnetic wave as a ray (in the optical sense), is often adequate. The deterministic model is used to evaluate the power of the received signal when there are no obstacles between the transmitter and receiver, that is in the presence of line of sight: in this case we can think of only one wave that propagates from the transmitter to the receiver. This situation is typical of transmissions between satellites and terrestrial radio stations in the microwave frequency range (3 < f 0 < 70 GHz). Let PT x be the power of the signal transmitted by an ideal isotropic antenna, which uniformly radiates in all directions in the free space. At a distance d from the antenna, the power density is 80 D PT x (W/m2 ) 4 d 2 (4.174)
where 4 d 2 is the surface of a sphere of radius d that is uniformly illuminated by the antenna. We observe that the power density decreases with the square of the distance. On a logarithmic scale (dB) this is equivalent to a decrease of 20 dB-per-decade with the distance. In the case of a directional antenna, the power density is concentrated within a cone and is given by 8 D G T x 80 D GT x PT x 4 d 2 (4.175)
where GT x is the transmit antenna gain. Obviously, GT x D 1 for an isotropic antenna; usually, GT x 1 for a directional antenna.
298
At the receive antenna, the available power in conditions of matched impedance is given by P Rc D 8A Rc Rc (4.176)
where P Rc is the received power, A Rc is the effective area of the receive antenna and Rc is the efciency of the receive antenna. The factor Rc < 1 takes into account the fact that the antenna does not capture all the incident radiation, because a part is reected or lost. To conclude, the power of the received signal is given by P Rc D PT x The antenna gain can be expressed as [1] 4 A (4.178) 2 where A is the effective area of the antenna, D c= f 0 is the wavelength of the transmitted signal, f 0 is the carrier frequency and is the efciency factor. The (4.178) holds for the transmit as well as for the receive antenna. We note that, because of the factor A=2 , working at higher frequencies presents the advantage of being able to use smaller antennas, for a given G. Usually 2 [0:5; 0:6] for parabolic antennas, while ' 0:8 for horn antennas. Observing (4.178), we get 2 (4.179) P Rc D PT x GT x G Rc 4 d GD The (4.179) is known as the Friis transmission equation and is valid in conditions of maximum transfer of power. The term .=4 d/2 is called free space path loss. Later, we will use the following denition: 2 P0 D PT x GT x G Rc (4.180) 4 which represents the power of a signal received at the distance of 1 meter from the transmitter. In any case, (4.179) does not take into account attenuation due to rain or other environmental factors, nor the possibility that the antennas may not be correctly positioned. The available attenuation of the medium, expressed in dB, is .ad /d B D 10 log10 PT x D 32:4 C 20 log10 djkm C 20 log10 f 0 jMHz P Rc .4=c/2 , .GT x /d B .G Rc /d B A Rc GT x Rc 4 d 2 (4.177)
where 32:4 D 10 log10 .G Rc /d B in dB. It is worthwhile making the following observations on the attenuation ad expressed by (4.181): a) it increases with distance as log10 d, whereas for metallic transmission lines the dependency is linear (see (4.140)); b) it increases with frequency as log10 f 0 . For GT x D G Rc D 1, .ad /d B coincides with the free space path loss.
299
We note that there are two noise sources, introduced by the antenna (w S ) and by the receiver (w A ). The noise temperature of the antenna depends on the direction in which the antenna is pointed: for example T S;Sun > T S;atmosphere (4.183)
Multipath
It is useful to study the propagation of a sinusoidal signal hypothesizing that the one-ray model is adequate, which implies using a directional antenna. Let sT x be a narrowband
300
transmitted signal, that is sT x .t/ D Re[A T x e j2 f 0 t ] The received signal at a distance d from the transmitter is given by s Rc .t/ D Re[A Rc e j2 f 0 .t
1 /
(4.184)
] D Re[A Rc e j' Rc e j2 f 0 t ]
(4.185)
where 1 D d=c denotes the propagation delay, A Rc is the amplitude of the received signal, and ' Rc D 2 f 0 1 D 2 f 0 d=c is the phase of the received signal. Using the denition (1.150) of h .a/ .t/, the radio channel associated with (4.185) has impulse response A Rc .a/ h . 1 / (4.186) gCh . / D Re AT x that is the channel attenuates the signal and introduces a delay equal to 1 . Choosing f 0 as the carrier frequency, the baseband equivalent of gCh is given by7
.bb/ gCh . / D
2A Rc e AT x
j2 f 0 1
1 /
(4.187)
Limited to signals sT x of the type (4.184), (4.186) can be rewritten as A Rc j' Rc .a/ e h . / gCh . / D Re AT x
(4.188)
Thus, (4.188) indicates that the received signal exhibits a phase shift of ' Rc D 2 f 0 1 with respect to the transmitted signal, because of the propagation delay. As the propagation delay is given by D d=c, the delay per unit of distance is equal to 3.3 ns/m. As the power decreases with the square of the distance between transmitter and receiver, the amplitude of the received signal decreases linearly with the distance, hence A Rc / A T x =d; in particular, if A0 is the amplitude of the received signal at the distance of 1 meter from the transmitter, then A Rc D A0 =d, and the power of the received signal is given by P Rc D A2 =2. Rc Reection and scattering phenomena imply that the one-ray model is applicable only to propagation in free space, and is not adequate to characterize radio channels, such as for example the channel between a xed radio station and a mobile receiver. We will now consider the propagation of a narrowband signal in the presence of reections. If a ray undergoes a reection caused by a surface, a part of its power is absorbed by the surface while the rest is re-transmitted in another direction. If the i-th ray has undergone K i reections before arriving at the receiver and if ai j is a complex number denoting the reection coefcient of the j-th reection of the i-th ray, the total reection factor is ai D
Ki Y jD1
ai j
(4.189)
.bb/
301
Therefore signal amplitudes, corresponding to rays that are not the direct or line of sight ray, undergo an attenuation due to reections that is added to the attenuation due to distance. The total phase shift asociated with each ray is obtained by summing the phase shifts introduced by the various reections and the phase shift due to the distance traveled. If Nc is the number of paths and di is the distance traveled by the i-th ray, extending the channel model (4.186) we get " # Nc A0 X ai .a/ gCh . / D Re h . i / (4.190) A T x i D1 di where i D di =c is the delay of the i-th ray. The complex envelope of the channel impulse response (4.190) around f 0 is equal to
.bb/ gCh . / D Nc 2A0 X ai e A T x i D1 di j2 f 0 i
i /
(4.191)
We note that the only difference between the passband model and its baseband equivalent is constituted by the additional phase term e j2 f 0 i for the i-th ray. Limited to narrowband signals, extending the channel model (4.188) to the case of many reections, the received signal can still be written as s Rc .t/ D Re[A Rc e j' Rc e j2 f 0 t ] where now amplitude and phase are given by A Rc e j' Rc D A0
Nc X ai e j'i di i D1
(4.192)
(4.193)
with 'i D 2 f 0 i . Let Ai and i be amplitude and phase, respectively, of the term A0 .ai =di /e j'i ; from (4.193) the resulting signal is given by the sum of Ai e j i , i D 1; : : : ; Nc , as represented in Figure 4.29. As P0 D A2 =2, the received power is 0
3 ARc R
c
302
P Rc
2 Nc X a i j'i D P0 e i D1 di
(4.194)
and is independent of the total phase of the rst ray. We will now give two examples of application of the previous results. Example 4.6.1 (Power attenuation as a function of distance in mobile radio channels) We consider two antennas, one transmitting and the other receiving, with height respectively h 1 and h 2 , that are placed at a distance d. Moreover, it is assumed that d h 1 and d h 2 (see Figure 4.30). We consider the case of two paths: one is the straight path (LOS), and the other is reected by the earth surface with reection coefcient a1 D 1, i.e. the earth acts as an ideal reecting surface and does not absorb power. Observing (4.194), and considering that for the above assumptions the lengths of the two paths are both approximately equal to d, the received power is given by P0 (4.195) P Rc ' 2 j1 e j1' j2 d where 1' D 2 f 0 1d=c D 2 1d= is the phase shift between the two paths, and 1d D 2h 1 h 2 =d is the difference between the lengths of the two paths. For small values of 1' we obtain: j1 e j1' j2 ' j1'j2 D 16 2 h2h2 1 2 2 d 2 (4.196)
from which, by substituting (4.180) in (4.195), we get h2h2 P0 j1'j2 D PT x GT x G Rc 1 4 2 (4.197) d2 d We note that the received power decreases as the fourth-power of the distance d, that is 40 dB/decade instead of 20 dB/decade as in the case of free space. Therefore the law of power attenuation as a function of distance changes in the presence of multipath with respect to the case of propagation in free space. P Rc D Example 4.6.2 (Fading caused by multipath) Consider again the previous example, but assume that transmitter and receiver are positioned in a room, so that the inequalities between the antenna heights and the distance d are no
LOS h1 h2
d
Figure 4.30. Two-ray propagation model.
303
longer valid. It is assumed, moreover, that the rays that reach the receive antenna are due, respectively, to LOS, reection from the oor, and reection from the ceiling. As a result the received power is given by 3 X a e j'i 2 i (4.198) P Rc D P0 i D1 di where the reection coefcients are a1 D 1 for the LOS path, and a2 D a3 D 0:7. With these assumptions, one nds that the power decreases with the distance in an erratic way, in the sense that by varying the position of the antennas the received power presents uctuations of about 2030 dB. In fact, depending on the position, the phases of the various rays change and the sum in (4.193) also varies: in some positions all rays are aligned in phase and the received power is high, whereas in others the rays cancel each other and the received power is low. In the previous example this phenomenon is not observed because the distance d is much larger than the antenna heights, and the phase difference between the two rays remains always small.
4.6.3
Doppler shift
In the presence of relative motion between transmitter and receiver, the frequency of the received signal undergoes a shift with respect to the frequency of the transmitted signal, known as a Doppler shift. We now analyze in detail the Doppler shift. With reference to Figure 4.31, we consider a transmitter radio Tx and a receiver radio that moves with speed v p from a point P to a point Q. The variation in distance between the transmitter and the receiver is 1` D v p 1t cos , where v p is the speed of the receiver relative to the transmitter, 1t is the time required for the receiver to go from P to Q, and is the angle of incidence of the signal with respect to the direction of motion ( is assumed to be the same in P and in Q). The phase variation of the received signal because of the different path length in P and Q is 1' D 2 v p 1t 2 1` D cos vp 1 1' D cos 2 1t (4.199)
(4.200)
Rc Q
304
This implies that if a narrowband signal given by (4.184) is transmitted, the received signal is s Rc .t/ D Re[A Rc e j2. f 0
f s /t
(4.201)
The (4.200) relates the Doppler shift to the speed of the receiver and the angle ; in particular, for D 0 we get f s D 9:259 10 4 v p jkm=h f 0 jMHz (Hz) (4.202) where v p km=h is the speed of the mobile in km/h, and f 0 jMHz is the carrier frequency in MHz. For example, if v p D 100 km/h and f 0 D 900 MHz we have f s D 83 Hz. We note that if the receiver moves towards the transmitter the Doppler shift is positive, if it moves away from the transmitter the Doppler shift is negative. We now consider a narrowband signal transmitted in an indoor environment8 where the signal received by the antenna is given by the contribution of many rays, each with a different length. If the signal propagation were taking place through only one ray, the received signal would undergo only one Doppler shift. But according to (4.200) the frequency shift f s depends on the angle . Therefore, because of the different paths, the received signal is no longer monochromatic, and we speak of a Doppler spectrum to indicate the spectrum of the received signal around f 0 . This phenomenon manifests itself also if both the transmitter and the receiver are static, but a person or an object moves modifying the signal propagation. The Doppler spectrum is characterized by the Doppler spread, which measures the dispersion in the frequency domain that is experienced by a transmitted sinusoidal signal. It is intuitive that the more the characteristics of the radio channel vary with time, the larger the Doppler spread will be. An important consequence of this observation is that the convergence time of algorithms used in receivers, e.g., to perform adaptive equalization, must be much smaller than the inverse of the Doppler spread of the channel, thus enabling the adaptive algorithms to follow the channel variations. Example 4.6.3 (Doppler shift) Consider a transmitter that radiates a sinusoidal carrier at the frequency of f 0 D 1850 MHz. For a vehicle traveling at 96.55 km/h (26.82 m/s), we want to evaluate the frequency of the received carrier if the vehicle is moving: a) approaching the transmitter, b) going away from the transmitter, c) perpendicular to the direction of arrival of the transmitted signal. The wavelength is 3 108 c D D 0:162 m f0 1850 106 a) The Doppler shift is positive; the received frequency is 26:82 f Rc D f 0 C f s D 1850 106 C D 1850:000166 MHz 0:162 D
8
(4.203)
(4.204)
The term indoor is usually referred to areas inside buildings, possibly separated by walls of various thickness, material, and height. The term outdoor, instead, is usually referred to areas outside of buildings: these environments can be of various types, for example, urban, suburban, rural, etc.
305
b) The Doppler shift is negative; the received frequency is 26:82 f Rc D f 0 f s D 1850 106 D 1849:999834 MHz 0:162 c) In this case cos. / D 0; therefore there is no Doppler shift.
(4.205)
4.6.4
For a wideband signal with spectrum centered around the carrier frequency f 0 , the channel model (4.191) is still valid; we rewrite the channel impulse response as a function of both the time variable t and the delay for a given t:
.bb/ gCh .t; / D Nc X i D1
gi .t/.
i .t//
(4.206)
where gi represents the complex-valued gain of the i-th ray that arrives with delay i . For a given receiver location, (4.206) models the channel as a linear lter having time-varying impulse response, where the channel variability is due to the motion of transmitter and/or receiver, or to changes in the surrounding environment, or to both factors. If the channel is time-invariant, or at least it is time-invariant within a short time interval, in this time interval the impulse response is only a function of . The transmitted signal undergoes three phenomena: a) fading of some gains gi due to multipath, which implies rapid changes of the received signal power over short distances (of the order of the carrier wavelength) and brief time intervals, b) time dispersion of the impulse response caused by diverse propagation delays of multipath rays, c) Doppler shift, which introduces a random frequency modulation that is in general different for different rays. In a digital transmission system the effect of multipath depends on the relative duration of the symbol period and the channel impulse response. If the duration of the channel impulse response is very small with respect to the duration of the symbol period, i.e. the transmitted signal is narrowband with respect to the channel, then the one-ray model is a suitable channel model; if the gain of the single ray varies in time we speak of a at fading channel. Otherwise, an adequate model must include several rays: in this case if the gains vary in time we speak of a frequency selective fading channel. Q Neglecting the absolute delay 1 .t/, letting 2 D 2 1 a simple two-ray radio channel model has impulse response
.bb/ gCh .t; / D g1 .t/. / C g2 .t/.
2 .t// Q
(4.207)
At a given instant t, the channel is equivalent to a lter with impulse response illustrated in Figure 4.32 and frequency response given by:9
.bb/ GCh .t; f / D g1 .t/ C g2 .t/e j2 f 2 .t/ Q
(4.209)
(4.208)
where b is a complex number. In the literature (4.208) is called Rummler model of the radio channel.
306
Figure 4.32. Physical representation and model of a two-ray radio channel, where g1 and g2 are assumed to be positive.
It is evident that the channel has a selective frequency behavior, as the attenuation depends on frequency. For g1 and g2 real-valued, from (4.209) the following frequency response is obtained 2 .bb/ 2 2 Q (4.210) GCh .t; f / D g1 .t/ C g2 .t/ C 2g1 .t/g2 .t/ cos.2 f 2 .t// shown in Figure 4.32. In any case, the signal distortion depends on the signal bandwidth in comparison to 1=2 . Q Going back to the general case, for wideband communications, rays with different delays are assumed to be independent, that is they do not interact with each other. In this case from (4.206) the received power is P Rc D PT x
Nc X i D1
jgi j2
(4.211)
From (4.211) we note that the received power is given by the sum of the squared amplitude of all the rays. Conversely, in the transmission of narrowband signals the received power is the square of the vector amplitude resulting from the vector sum of all the received rays. Therefore, for a given transmitted power, the received power will be lower for a narrowband signal as compared to a wideband signal.
307
q 2
(4.212)
jgi j2 in n D 1; 2 jgi j
2
n D
i D1 Nc X i D1
(4.213)
The above formulae give the rms delay spread for an instantaneous channel impulse response. With reference to the time-varying characteristics of the channels, we use the (average) rms delay spread r ms obtained by substituting in (4.213) in place of jgi j2 its expectation. In this case r ms measures the mean time dispersion that a signal undergoes because of multipath. Typical values of (average) rms delay spread are of the order of s in outdoor mobile radio channels, and of the order of some tenths of ns in indoor channels. We dene as power delay prole, also called delay power spectrum or multipath intensity prole, the expectation of the squared amplitude of the channel impulse response, E[jgi j2 ], as a function of delay i . In Table 4.5 power delay proles are given for some typical channels.
where C is a real-valued constant and gi is a complex-valued random variable with zero Q mean and Gaussian distribution (see Example 1.9.3 on page 67). In other words, whereas the rst ray contains a direct (deterministic) component in addition to a random component, all the other rays are assumed to have only a random component: therefore the distribution
308
Table 4.5 Values of E[jgi j2 ] (in dB) and i (in ns) for three typical channels.
Standard GSM i 0 200 500 1600 2300 5000 E[jgi j2 ] 3:0 0 2:0 6:0 8:0 10:0
Indoor ofces i 0 50 150 325 550 700 E[jgi j2 ] 0:0 1:6 4:7 10:1 17:1 21:7
Indoor business i 0 50 150 225 400 525 750 E[jgi j2 ] 4:6 0 4:3 6:5 3:0 15:2 21:7
of jgi j will be a Rice distribution for jg1 j and a Rayleigh distribution for jgi j, i 6D 1. In p particular, letting gi D gi = E[jgi j2 ], we have N p pjg1 j .a/ D 2.1 C K / a exp[ K .1 C K /a 2 ]I0 [2a K .1 C K /]1.a/ N (4.215) 2 pjgi j .a/ D 2a e a 1.a/ N where I0 is the modied Bessel function of the rst type and order zero, Z 1 e x cos d I0 .x/ D 2
(4.216)
The probability density (4.215) is given in Figure 4.33 for various values of K . In (4.214) the phase of gi is uniformly distributed in [0; 2 /. For a one-ray channel Q Q model, the parameter K D C 2 =E[jg1 j2 ], known as the Rice factor, is equal to the ratio between the power of the direct component and the power of the reected and/or scattered component. In general for a model with more rays we take K D C 2 =Md , where Md is the P statistical power of all reected and/or scattered components, that is Md D iNc E[jgi j2 ]. Q D1 Assuming that the power delay prole is normalized such that
Nc X i D1
E[jgi j2 ] D 1
(4.217)
p we obtain C D K =.K C 1/. Typical reference values for K are 3 and 10 dB. If C D 0, i.e. no direct component exists, it is K D 0, and the Rayleigh distribution is obtained for all the gains fgi g. For K ! 1, i.e. with no reected and/or scattered components and, hence, C D 1, we nd the model having only the deterministic component. To justify the Rice model for jg1 j we consider the transmission of a sinusoidal signal (4.184). In this case the expression of the received signal is given by (4.192), which we rewrite as follows: Q s Rc .t/ D [g1;I .t/ C C] cos 2 f 0 t g1;Q .t/ sin 2 f 0 t Q (4.218)
309
2 K=10
1.8
1.6
K=5
1.4
1.2
(a)
K=2
1
1
|g |
K=0
0.8
0.6
0.4
0.2
0.5
1.5 a
2.5
Figure 4.33. The Rice probability density function for various values of K. The Rayleigh density function is obtained for K D 0.
where C represents the contribution of the possible direct component of the propagation Q Q Q signal, and g1;I and g1;Q are due to the scattered component. As the gains g1;I and g1;Q Q are given by the sum of a large number of random components, they can be approximated by independent Gaussian random processes with zero mean. The instantaneous envelope of the received signal is then given by q [g1;I .t/ C C]2 C g1;Q .t/ Q Q2 (4.219) which, in the assumption just formulated, is a Rice random variable for each instant t.
4.6.5
The channel model previously studied is especially useful for system simulations, as will be discussed later. A general continuous-time model is now presented. Assuming that the signal propagation occurs through a large number of paths, which in turn are subject to a very large number of random phenomena, the (baseband equivalent) channel impulse response can be represented with good approximation as a time-varying complex-valued Gaussian random process g.t; /. In particular g.t; / represents the channel output at the instant t in response to an impulse applied at the instant .t /. We now evaluate the autocorrelation function of the impulse response evaluated at two different instants and two different delays, rg .t; t 1t;; 1 / D E[g.t; /g .t 1t; 1 /] (4.220)
310
According to the model known as the wide-sense stationary uncorrelated scattering (WSSUS), the values of g for rays that arrive with different delays are uncorrelated, and g is stationary in t. Therefore we have rg .t; t 1t;; 1 / D rg .1t; /.1 / (4.221)
In other words, the autocorrelation is non-zero only for impulse responses that are considered for the same delay time. Moreover, as g is stationary in t, if the delay time is the same, the autocorrelation only depends on the difference of the times at which the two impulse responses are evaluated.
2 r ms /
(4.222)
2 1 e r ms
2 2 =.2 r ms /
(4.223)
(4.224)
The measure of the set of values for which M. / is above a certain threshold is called (average) excess delay spread of the channel. As in the case of the discrete channel model previously studied, we dene the (average) rms delay spread as Z 1 . /2 M. / d 1 2 Z 1 (4.225) r ms D M. / d
1
where D Z
1 1
M. / d
(4.226)
The inverse of the (average) rms delay spread is called the coherence bandwidth of the channel.
311
For digital transmission over such channels, we observe that if r ms is of the order of 20% of the symbol period, or larger then signal distortion is non-negligible. Equivalently, if the coherence bandwidth of the channel is lower than 5 times the modulation rate of the transmission system, then we speak of a frequency selective fading channel, otherwise the channel is at fading. However, in the presence of at fading the received signal may vanish completely, whereas frequency selective fading produces several replicas of the transmitted signal at the receiver, so that a suitably designed receiver can recover the transmitted information. Example 4.6.4 (Power delay prole) We compute the average rms delay spread for the multipath delay prole of Figure 4.34, 1 and determine the coherence bandwidth, dened as Bc D 5r ms . From (4.213) we have D N and .1/.5/2 C .0:1/.1/2 C .0:1/.2/2 C .0:01/.0/ D 21:07 . s/2 N2 D 0:01 C 0:1 C 0:1 C 1 Therefore we get r ms D N p 21:07 .4:38/2 D 1:37 s (4.229) (4.228) .1/.5/ C .0:1/.1/ C .0:1/.2/ C .0:01/.0/ D 4:38 s 0:01 C 0:1 C 0:1 C 1 (4.227)
Doppler spectrum
We now analyze the WSSUS channel model with reference to time variations. First we introduce the correlation function of the channel frequency response taken at instants t and t 1t, and, respectively, at frequencies f and f 1 f , rG .t; t 1t; f; f 1 f / D E[G.t; f /G .t 1t; f 1 f /] (4.230)
M( ) (dB)
312
g.t; / e
j2 f
(4.231)
(4.232)
that is rG .1t;1 f / is the Fourier transform of rg .1t; /. The Fourier transform of rG is given by Z 1 PG .; 1 f / D rG .1t;1 f /e j2 .1t/ d.1t/ (4.233)
1
The time variation of the frequency response is measured by PG .; 0/. Now we introduce the Doppler spectrum D./, which represents the power of the Doppler shift for different values of the frequency . We recall that the Doppler shift is caused by the motion of terminals or surrounding objects. We dene D./ as the Fourier transform of the autocorrelation function of the impulse response, in correspondence of the same delay , evaluated at two different instants, that is:10 Z C1 rg .1t; / j2 1t e D./ D d.1t/ (4.234) 1 rg .0; / The term rg .0; / in (4.234) represents a normalization factor such that Z C1 D./ d D 1
1
(4.235)
We note that (4.234) implies that rg .1t; / is a separable function, rg .1t; / D d.1t/ rg .0; / D d.1t/ M. / where d.1t/ D F
1 [D./],
(4.236)
(4.238)
With the above assumptions the following equality holds: D./ D PG .;0/ (4.239)
10 In very general terms, we could have a different Doppler spectrum for each path, or gain g.t; /, of the channel.
313
Therefore, D./ can also be obtained as the Fourier transform of rG .1t;0/, which in turn can be determined by transmitting a sinusoidal signal (hence 1 f D 0) and estimating the autocorrelation function of the amplitude of the received signal. The maximum frequency f d of the Doppler spectrum support is called the Doppler spread of the channel and gives a measure of the fading rate of the channel. Another measure of the support of D./ can be obtained through the rms Doppler spread or second order central moment of the Doppler spectrum. The inverse of the Doppler spread is called coherence time: it gives a measure of the time interval within which a channel can be assumed to be time invariant or static. Let T be the symbol period in a digital transmission system; we usually say that the channel is fast fading if f d T > 10 2 , and slow fading if f d T < 10 3 .
J0 .2 f d 1t/ M.i / .
i /
(4.241)
where J0 is the Bessel function of the rst type and order zero. The model of the Doppler spectrum described above agrees with the experimental results obtained for mobile radio channels. For indoor radio channels, thanks to the study conducted by a special commission (JTC), it was demonstrated that the Doppler spectrum can be modelled as 8 < 1 j f j fd (4.242) D. f / D 2 f d : 0 elsewhere with a corresponding autocorrelation function given by rg .1t; / D
Nc X i D1
i /
(4.243)
A further model assumes that the Doppler spectrum is described by a second or third-order Butterworth lter with the 3 dB cutoff frequency equal to f d .
Shadowing
The simplest relation between average transmitted power and average received power is P Rc D P0 d (4.244)
314
where is equal to 2 for propagation in free space and to 4 for the simple 2-ray model described before. For indoor and urban outdoor radio channels the relation depends on the environment, according to the number of buildings, their dimensions, and also the material used for their construction; in general, however, variations of the average received power are lower in outdoor environments than in indoor environments. Shadowing takes into account the fact that the average received power may present uctuations around the value obtained by deterministic models. These uctuations are modelled as a log-normal random variable, that is e , where is a Gaussian random variable with zero mean and variance 2 . If P Rc is the average received power obtained by deterministic rules, in the presence of shadowing it becomes e P Rc ; in practice shadowing provides a measure of the adequacy of the adopted deterministic model. A propagation model that completely ignores any information on land conguration, and therefore is based only on the distance between transmitter and receiver, has a shadowing with . /d B D 12 dB. The relation between and . /d B is D 0:23. /d B . Improving the accuracy of the propagation model, for example, by using more details regarding the environmental conguration, the shadowing can be reduced; in case we had an enormous amount of topographic data and the means to elaborate them, we would have a model with D 0. Hence, shadowing should be considered in the performance evaluation of mobile radio systems, whereas for the correct design of a network it is good practice to make use of the largest possible quantity of topographic data.
Final remarks
A signal that propagates in a radio channel for mobile communications undergoes a type of fading that depends on the signal as well as on the channel characteristics. In particular, whereas the delay spread due to multipath leads to dispersion in the time domain and therefore frequency selective fading, the Doppler spread causes dispersion in the domain of the variable and therefore time selective fading. The rst type of fading can be divided into at fading and frequency selective fading. In the rst case the channel has a constant gain; in other words, the inverse of the transmitted signal bandwidth is much larger than the delay spread of the channel and g.t; / can be approximated by a delta function, with random amplitude and phase, centered at D 0. In the second case instead the channel has a time-varying frequency response within the passband of the transmitted signal and consequently the signal undergoes frequency selective fading; these conditions occur when the inverse of the transmitted signal bandwidth is of the same order or smaller than the delay spread of the channel. The received signal consists of several attenuated and delayed versions of the transmitted signal. A channel can be fast fading or slow fading. In a fast fading channel, the impulse response of the channel changes within a symbol period, that is the coherence time of the channel is smaller than the symbol period; this condition leads to signal distortion, which increases with increasing Doppler spread. Usually there are no remedies to compensate for such distortion unless the symbol period is decreased; on the other hand, this choice leads to larger intersymbol interference. In a slow fading channel, the impulse response changes much more slowly with respect to the symbol period. In general, the channel can be assumed as time invariant for a time interval that is proportional to the inverse of the Doppler spread.
315
4.6.6
Our aim is to approximate a transmission channel dened in the continuous-time domain by a channel in the discrete-time domain characterized by sampling period TQ . We immediately notice that the various delays in (4.206) must be multiples of TQ and consequently we need to approximate the delays of the power delay prole (see, e.g., Table 4.5). Starting from a continuous-time model of the power delay prole (see, e.g., (4.224)), we need to obtain a sampled version of M. /. The discrete-time model of the radio channel is represented, as illustrated in Figure 4.35, by a time-varying linear lter where the coefcient gi corresponds to the complex gain of the ray with delay i TQ , i D 0; 1; : : : ; Nc 1; in the case of at fading we choose Nc D 1. If the channel is time invariant ( f d D 0), all coefcients fgi g, i D 0; : : : ; Nc 1, are constant, and are obtained as realizations of Nc random variables. In general, however, fgi g are random processes. To generate each process gi .kTQ /, the scheme of Figure 4.36 is used, where wi .`T P / is complex-valued Gaussian white noise with zero N mean and unit variance, h ds is a narrrowband lter that produces a signal gi0 with the desired Doppler spectrum, and h int is an interpolator lter (see Section 1.A.7). Usually we choose f d TQ 1, and 1=10 p f d TP 1=5. The interpolator output signal is then multiplied by a constant i D M.i TQ /, which imposes the desired power delay prole.
x(kTQ) TQ TQ TQ
g (kTQ)
0
g (kTQ)
1
g (kT ) Q
N-1 c
+ y(kTQ)
Figure 4.35. Discrete time model of a radio channel.
316
If the channel model includes a deterministic component for the ray with delay i D i TQ , a constant Ci must be added to the random component gi . Furthermore, if the channel model N N includes a Doppler shift f si for the i-th branch, then we need to multiply the term Ci C gi by the exponential function exp. j2 f si kTQ /. Observing (4.211), to avoid modifying the average transmitted power, the coefcients fgi g, i D 0; 1; : : : ; Nc 1, are scaled, so that
Nc 1 X i D0
(4.245)
For example, the above condition is satised if each signal gi0 has unit statistical power11 and fi g satisfy the condition
Nc 1 X i D0
.i2 C Ci2 / D 1
(4.246)
where, dening !0 D tan.!d TP =2/ where TP is the sampling period, we have [18] a1 D
2 !0 / p 2 1 C !0 C 2 !0
2.1
(4.248)
a2 D c0 D
4 1 C !0 p 2 .1 C !0 C 2 !0 /2 1 4
(4.249) (4.250)
.1 C a1 C a2 /
11 Based on the Example 1.9.10 on page 72, it is M 0 D 1 if M wi D 1; the equivalent interpolator lter, given by N gi
the cascade of h ds and h int , has energy equal to the interpolation factor T P =TQ .
317
The lter output gives gi0 .`T P / D a1 gi0 ..` 1/T P / a2 gi0 ..` 2/T P / 2/T P // (4.251)
1/T P / C wi ..` Q
1.b) IIR lter with classical Doppler spectrum. Now h ds is implemented as the cascade of two lters. The rst, Hds 1 .z/, is an FIR shaping lter with amplitude characteristic of the frequency response given by the square root of the function in (4.240). The second, Hds 2 .z/, is a Chebychev lowpass lter, with cutoff frequency f d . Table 4.6 reports values of the overall lter parameters for f d TP D 0:1 [19]. 2) Using sinusoidal signals. Let gi0 .`T P / D
Nf X mD1
j .2 f m `TP C'i;m /
j8i;Q
(4.252)
The spacing between the different frequencies is 1 f m ; letting f 1 D 1 f 1 =2, for m > 1 we have f m D f m 1 C 1 f m . Each 1 f m can be chosen as a constant, 1 fm D Z or, dening K d D
0 fd
fd Nf
(4.253)
Table 4.6 Parameters of an IIR lter which implements a classical Doppler spectrum. [From Anastasopoulos and Chugg (1997). c 1997 IEEE.]
fan g; n D 0; : : : ; 11 : 1:0000 e C 0 4:4153 e C 0 6:1051 e C 0 1:3542 e C 0 7:9361 e C 0 5:1221 e C 0 fbn g; n D 0; : : : ; 21 : 1:3651 e 4 8:1905 e 2:0476 e 3 9:0939 e 1:8067 e 3 1:3550 e 7:1294 e 5 9:5058 e 1:3321 e 5 4:5186 e 1:8074 e 5 3:0124 e
4 4 3 5 5 6
3 4 4 5 5
3 3 5 5 5
318
Figure 4.37. Nine realizations of jgCh .t; /j for a Rayleigh channel with an exponential power delay prole having rms D 0:5 T.
(bb)
Nf XZ mD1
. fm
f /2 D. f / d f
(4.255)
The phases 'i;m , 8i;I and 8i;Q are uniformly distributed in [0; 2 / and statistically independent. This choice for 8i;I and 8i;Q ensures that the real and imaginary parts of gi0 are statistically independent. p The amplitude is given by Ai;m D D. f m /1 f m . If D. f / is at, by the central limit theorem we can claim that gi0 is a Gaussian process; if instead D. f / presents some frequencies with large amplitude, Ai;m must be generated as a Gaussian random variable with zero mean and variance D. f m /1 f m . In Figure 4.37 are represented nine realizations of the amplitude of the impulse response of a Rayleigh channel obtained by the simulation model of Figure 4.35, for an exponential power delay prole with r ms D 0:5 T . The Doppler frequency f d was assumed to be zero. We point out that the parameter r ms provides scarce information on the actual behavior .bb/ of gCh , which can scatter for a duration equal to 4-5 times r ms .
4.7
4.7.1
Telephone channel
Characteristics
Telephone channels, originally conceived for the transmission of voice, today are extensively used also for the transmission of data. Transmission of a signal over a telephone
319
channel is achieved by utilizing several transmission media, such as symmetrical transmission lines, coaxial cables, optical bers, radio, and satellite links. Therefore channel characteristics depend on the particular connection established. As a statistical analysis made in 1983 indicated [2], a telephone channel is characterized by the following disturbances and distortions.
Linear distortion
The frequency response GCh . f / of a telephone channel can be approximated by a passband lter with band in the range of frequencies from 300 to 3400 Hz. The plots of the attenuation a. f / D 20 log10 jGCh . f /j (4.256)
and of the group delay or envelope delay (see (1.149)) . f / D 1 d arg GCh . f / 2 d f (4.257)
are illustrated in Figure 4.38 for two typical channels. The attenuation and envelope delay distortion are normalized by the values obtained for f D 1004 Hz and f D 1704 Hz, respectively.
Noise sources
Impulse noise. It is caused by electromechanical switching devices and is measured by the number of times the noise level exceeds a certain threshold per unit of time. Quantization noise. It is introduced by the digital representation of voice signals and is the dominant noise in telephone channels (see Chapter 5). For a single quantizer, the signal-to-quantization noise ratio 3q has the behavior illustrated in Figure 4.39. Thermal noise. It is described in Section 4.2 and is present at a level of 20 30 dB below the desired signal.
Non-linear distortion
It is caused by ampliers and by non-linear A-law and -law converters (see Chapter 5).
Frequency offset
It is caused by the use of carriers for frequency up and downconversion. The relation between the channel input x.t/ and output y.t/ is given by ( X. f f off / f >0 Y. f/ D (4.258) X . f C f off / f <0 Usually f off 5 Hz.
320
Figure 4.38. Attenuation and envelope delay distortion for two typical telephone channels.
321
Figure 4.39. Signal to quantization noise ratio as a function of the input signal power for three different inputs.
talker echo
listener echo
Figure 4.40. Three of the many signal paths in a simplied telephone channel with a single two-to-four wire conversion at each end.
Phase jitter
It is a generalization of the frequency offset (see (4.270)).
Echo
As discussed in Section 3.6.5, it is caused by the mismatched impedances of the hybrid. As illustrated in Figure 4.40, there are two types of echoes: 1. Talker echo: part of the signal is reected and input to the receiver at the transmit side. If the echo is not very delayed, then it is practically indistinguishable from the original voice signal;
322
2. Listener echo: if the echo is reected a second time, it returns to the listener and disturbs the original signal. On terrestrial channels the round-trip delay of echoes is of the order of 1060 ms, whereas on satellite links it may be as large as 600 ms. We note that the effect of echo is similar to multipath fading in radio systems. To mitigate the effect of echo there are two strategies: use echo suppressors that attenuate the unused connection of a four-wire transmission line; use echo cancellers that cancel the echo at the source, as illustrated in the scheme of Figure 3.36.
4.8
In this section we will describe a transmission channel model that takes into account the non-linear effects due to the transmitter and the disturbance introduced by the receiver and by the channel. We will now analyze the various blocks of the baseband equivalent model illustrated in Figure 4.41.
where A.t/ 0 is the signal envelope, and '.t/ is the instantaneous phase deviation.
Figure 4.41. Baseband equivalent model of a transmission channel including a non-linear device.
323
The envelope and the phase of the output signal, sT x .t/, depend on the instantaneous, i.e. without memory, transformations of the input: sT x .t/ D G[A.t/] cos.2 f 0 t C '.t/ C 8[A.t/]/ It is usually more convenient to refer to baseband equivalent signals: s .bb/ .t/ D A.t/e j'.t/ and
.bb/ sT x .t/ D G[A.t/]e j .'.t/C8[A.t/]/
(4.260)
(4.261)
(4.262)
The functions G[A] and 8[A], called envelope transfer functions, represent respectively the amplitude/amplitude (AM/AM) conversion and the amplitude/phase (AM/PM) conversion of the amplier. In practice, the HPA are of two types. For each type we give the AM/AM and AM/PM functions commonly adopted for the analysis. First, however, we need to introduce some normalizations. As a rule, the point at which the amplier operates is identied by the back-off. We adopt here the following denitions for the input back-off (IBO) and the output back-off (OBO): S (dB) (4.263) IBO D 20 log10 p Ms ST x (dB) (4.264) OBO D 20 log10 p Ms T x where Ms is the statistical power of the input signal s, MsT x is the statistical power of the output signal sT x , and S and ST x are the amplitudes of the input and output signals, respectively, that lead to saturation of the amplier. Here we assume S D 1 and ST x D G[1] for all the ampliers considered. TWT. The travelling wave tube (TWT) is a device characterized by a strong AM/PM conversion. The conversion functions are G[A] D 8[A] D A A 1 C A A2 8 A 2 1 C 8 A 2 (4.265)
(4.266)
where A , A , 8 and 8 suitable parameters. The (4.265) and (4.266) are illustrated in Figure 4.42 for A D 1, A D 0:25, 8 D 0:26 and 8 D 0:25. SSPA. The solid state power amplier (SSPA) has a more linear behavior in the region of small signals as compared to the TWT. The AM/PM conversion is usually negligible.
324
G[A] (dB)
10
15
14
12
10
6 A (dB)
(a)
25
20
15
[A] (deg.)
10 5 0
14
12
10
6 A (dB)
(b)
Figure 4.42. AM/AM and AM/PM characteristics of a TWT for A D 1, A D 0:25, 8 D 0:26 and 8 D 0:25.
8[A] D 0
where p is a suitable parameter. In Figure 4.43 the function G[A] is plotted for three values of p; the superimposed dashed line is an ideal curve given by ( A 0< A<1 G[A] D (4.269) 1 A1
325
G[A] (dB)
10
15
14
12
10
6 A (dB)
5
HPA 38 GHz HPA 40 GHz
G[A] (dB)
10
15
14
12
10
6 A (dB)
Figure 4.44. AM/AM experimental characteristic of two ampliers operating at 38 GHz and 40 GHz.
326
It is interesting to compare the above analytical models with the behavior of a practical HPA. Figure 4.44 illustrates the AM/AM characteristics of two waveguide HPA operating at frequencies of 38 GHz and 40 GHz.
Transmission medium
The transmission medium is typically modelled as a lter. For transmission lines and radio links the models are given respectively in Sections 4.4 and 4.6.
Additive noise
Several noise sources that cause a degradation of the received signal may be present in a transmission system. Consider, for example, the noise introduced by a receive antenna or the thermal noise and shot noise generated by the pre-amplier stage of a receiver. At the receiver input, all these noise signals are modelled as an effective additive white Gaussian noise (AWGN) signal, statistically independent of the desired signal. The power spectral density of the AWGN noise can be obtained by the analysis of the system devices, or by experimental measurements.
Phase noise
The demodulators used at the receivers are classied as coherent or non-coherent, depending on whether they use or not a carrier signal, which ideally should have the same phase and frequency as the carrier at the transmitter, to demodulate the received signal. Typically both phase and frequency are recovered from the received signal by a phase locked loop (PLL) system, which employs a local oscillator. The recovered carrier may differ from the transmitted carrier because of the phase noise,12 due to short-term stability, i.e. frequency drift, of the oscillator, and because of the dynamics and transient behavior of the PLL. The recovered carrier is expressed as dt 2 (4.270) v.t/ D Vo [1 C a.t/] cos !0 t C ' j .t/ C 2 where d (long-term drift) represents the effect due to ageing of the oscillator, a.t/ is the amplitude noise, and ' j .t/ denotes the phase noise. Often the amplitude noise a.t/, as well as the effect of ageing, can be neglected. The phase noise is usually represented in a transmission system model as in Figure 4.41. The phase noise ' j .t/ consist of deterministic components and random noise. For example, temperature change, supply voltage, and the output impedance of the oscillator are included among deterministic components.
327
Ignoring the deterministic effects, with the exception of the frequency drift, a PSD model of the of ' j .t/ comprises ve terms: f2 f2 P' j . f / D k 4 0 C k 3 0 C f4 f3 | {z } | {z }
random frequency walk icker frequency noise
f2 k 2 0 f2 | {z }
f2 C k 1 0 C f | {z }
icker phase noise
k0 f 2 | {z0 }
(4.271)
for f ` f f h . A simplied model, often used, is given by 8 >a < P' j . f / D c C >b 1 : f2
j f j f1 f1 j f j < f2 (4.272)
where the parameters a and c are typically of the order of 65 dBc/Hz and 125 dBc/Hz, respectively, and b is a scaling factor that depends on f 1 and f 2 and assures continuity of the PSD. dBc means dB carrier, that is it represents the statistical power of the phase noise, expressed in dB, with respect to the statistical power of the desired signal received in the passband. Depending on the values of a, b, c, f 1 and f 2 , typical values of the statistical power of ' j .t/ are in the range from 10 2 to 10 4 . The plot of (4.272) is shown in Figure 4.45 for f 1 D 0:1 MHz, f 2 D 2 MHz, a D 65 dBc/Hz, and c D 125 dBc/Hz.
60
70
80
P(f) (dBc/Hz)
90
100
110
(~ 20 dB/decade)
120
130
10
15
f (MHz)
328
Bibliography
[1] C. G. Someda, Electromagnetic waves. London: Chapman & Hall, 1998. [2] M. of the Technical Staff, Transmission systems for communications. Winston, NC: Bell Telephone Laboratories, 5th ed., 1982. [3] S. Ramo, J. R. Whinnery, and T. Van Duzer, Fields and Waves in Communication Electronics. New York: John Wiley & Sons, 1965. [4] G. Cherubini, S. Olcer, G. Ungerboeck, J. Creigh, and S. K. Rao, 100BASE-T2: a new standard for 100 Mb/s ethernet transmission over voice-grade cables, IEEE Communications Magazine, vol. 35, pp. 115122, Nov. 1997. [5] R. J. Hoss, Fiber optic communications. Englewood Cliffs, NJ: Prentice-Hall, 1990. [6] L. B. Jeunhomme, Single-mode ber optics. New York: Marcel Dekker, 2nd ed., 1990. [7] J. C. Palais, Fiber optic communications. Englewood Cliffs, NJ: Prentice-Hall, 3rd ed., 1992. [8] T. Li, Structures, parameters, and transmission properties of optical bers, Proc. IEEE, p. 1175, Oct. 1980. [9] T. Miya, Y. Terunuma, T. Hosaka, and T. Miyashita, Ultimate low-loss single-mode bre at 1.55 m, IEE Electronics Letters, vol. 15, pp. 106108, Feb. 1979. [10] J. C. Palais, Fiber optic communications systems, in The Communications Handbook (J. D. Gibson, ed.), ch. 54, pp. 731739, Boca Raton: CRC Press, 1997. [11] D. G. Messerschmitt and E. A. Lee, Digital communication. Boston, MA: Kluwer Academic Publishers, 2nd ed., 1994. [12] K. Feher, Wireless digital communications. Upper Saddle River, NJ: Prentice-Hall, 1995. [13] T. S. Rappaport, Wireless communications: principles and practice. Englewood Cliffs, NJ: Prentice-Hall, 1996. [14] K. Pahlavan and A. H. Levesque, Wireless information networks. New York: John Wiley & Sons, 1995. [15] W. C. Jakes, Microwave mobile communications. New York: IEEE Press, 1993. [16] G. L. Stuber, Principles of mobile communication. Norwell, MA: Kluwer Academic Publishers, 1996. [17] J. J. Spilker, Digital communications by satellite. Englewood Cliffs, NJ: Prentice-Hall, 1977.
4. Bibliography
329
[18] A. V. Oppenheim and R. W. Schafer, Discrete-time signal processing. Englewood Cliffs, NJ: Prentice-Hall, 1989. [19] A. Anastasopoulos and K. Chugg, An efcient method for simulation of frequency selective isotropic Rayleigh fading, in Proc. 1997 IEEE Vehicular Technology Conference, pp. 20842088, May 1997.
Chapter 5
Figure 5.1a illustrates the conventional transmission of an analog signal, for example, speech or video, over an analog channel; in this scheme the transmitter usually consists of an amplier and possibly a modulator, the analog transmission channel is of the type discussed in Chapter 4, and the receiver consists of an amplier and possibly a demodulator. Alternatively, the transmission may take place by rst encoding1 the information contained in the analog signal into a sequence of bits using for example an analog-to-digital converter (ADC), as illustrated in Figure 5.1b. If Tb is the time interval between two consecutive bits of the sequence, the bit rate of the ADC is Rb D 1=Tb (bit/s). The binary message is converted by a digital modulator into a waveform that is suitable for transmission over an analog channel. At the receiver, the reverse process occurs: in this case a digital demodulator restores the message, whereas the conversion of the sequence of bits to an analog signal is performed by a digital-to-analog converter (DAC). The system that has as an input the sequence of bits produced by the ADC, and as an output the sequence of bits produced by the digital demodulator is called a binary channel (see Chapter 7). In this chapter the principles and methods for the conversion of analog signals into binary messages and viceversa will be discussed; as a practical example we will use speech, but the principles may be extended to any analog signal. To compute system performance, a fundamental parameter is the signal-to-noise ratio. Let s.t/ be the original signal, s .t/ the Q Q reconstructed signal, and eq .t/ D s .t/ s.t/; then the signal-to-noise ratio is dened as 3q D E[s 2 .t/] 2 E[eq .t/] (5.1)
5.1
Analog access over a telephone channel in the public switched telephone network (PSTN) is illustrated in Figure 5.2. With reference to the gure, the word modem is the contraction of mod ulator-demodulator. Its function is to convert a binary message, or data signal, into
1
We bring to the attention of the reader that the terms encoder and decoder are commonly used to indicate various devices in a communication system. In this chapter we will deal with encoders and decoders for the digital representation of analog waveforms.
332
an analog passband signal that can be transmitted over the telephone channel. In Figure 5.2, the source generates a speech signal or a data le; in the latter case, a modem is required to transmit the signal. The analog signal s.t/ that has a band of approximately 3003400 Hz is sent over a local loop to the central ofce (see Chapter 4): here it is usually converted into a binary digital message via PCM at 64 kbit/s; in turn this message is modulated before being transmitted over an analog channel. After having crossed several central ofces where switching (routing) of the signal takes place, the PCM encoded message arrives at the destination central ofce: here it is converted into an analog signal and sent over a local loop to the end user. It is here that the signal must be identied as a speech signal or a digitally modulated signal; in the latter case a modem will demodulate it to reproduce the data message. Figure 5.3 illustrates the concept of direct digital access at the users premises. An analog signal is converted into a digital message via an ADC. The user digital message is then sent over the analog channel by a modulator. At the receiver the inverse process is established, where the digital message obtained at the output of the demodulator may be used to restore an analog signal via a DAC. In comparing the two systems, we note the waste of capacity of the system in Figure 5.2. For example, for a 9600 bit/s modem, the modulated PCM encoded signal requires a standard capacity of Rb D 64 kbit/s. By directly accessing the PCM link at the users home, we could transmit 64000=9600 ' 6 data signals at 9600 bit/s.
5.1.1
Some waveforms
Some examples of speech waveforms for an interval of 0.25 s are given in Figure 5.4. From these plots, we can obtain a speech model as a succession of voiced speech spurts (see Figure 5.5a), or unvoiced speech spurts (see Figure 5.5b). In the rst case, the signal
333
334
335
is strongly correlated and almost periodic, with a period that is called pitch, and exhibits large amplitudes; conversely in an unvoiced speech spurt the signal is weakly correlated and has small amplitudes. We note moreover that the average level of speech changes in time: indeed speech is a non-stationary signal. In Figure 5.6 it is interesting to observe the instantaneous spectrum of some voiced and unvoiced sounds; we also note that the latter may have a bandwidth larger than 10 kHz. Concerning the amplitude distribution of speech signals, we observe that over short time intervals, of the order of a few tenths of milliseconds (or of a few hundreds of samples at a sampling frequency of 8 kHz), the amplitude statistic is Gaussian with good approximation; over long time intervals, because of the numerous pauses in speech, it tends to exhibit a gamma or Laplacian distribution. We give here the probability density functions of the amplitude that are usually adopted. Let s be the standard deviation of the signal s.t/; then we have gamma: Laplacian: Gaussian: ps .a/ D 3 8 s jaj 1 p !1
2 p
e
2jaj s 1 a 2 s 2
3jaj 2s
ps .a/ D p e 2s ps .a/ D p e 2 s 1
(5.2)
As mentioned above, analog modulated signals generated by modems, often called voiceband data signals, are also transmitted over telephone channels. Figure 5.7 illustrates a
Figure 5.6. Spectrum of voiced and unvoiced sounds for a sampling frequency of 20 kHz.
336
s(t)
0.06
t (s)
Figure 5.7. Signal generated by a modem employing FSK modulation for the transmission of 1200 bit/s.
Figure 5.8. Signal generated by modems employing PSK modulation for the transmission of: (a) 2400 bit/s; (b) 4800 bit/s.
signal produced by the 202S modem, which employs FSK modulation for the transmission of 1200 bit/s, whereas Figure 5.8a and Figure 5.8b illustrate signals generated by the 201C and 208B modems, which employ PSK modulation for the transmission of 2400 and 4800 bit/s, respectively. For the denition of FSK and PSK modulation we refer the reader to Chapter 6. In general, we note that the average level of signals generated by modems is stationary; moreover, if the bit rate is low, signals are strongly correlated.
337
Speech coding
Speech coding addresses person-to-person communications and is strictly related to the transmission, for example, over the public network, and storage of speech signals. The aim is to represent, using an encoder, speech as a digital signal that requires the lowest possible bit rate to recreate, by an appropriate decoder, the speech signal at the receiver [1]. Depicted in Figure 5.9 is a basic scheme, denoted as ADC, that provides the analog-todigital conversion (encoding) of the signal, consisting of: 1. an anti-aliasing lter followed by a sampler at sampling frequency 1=Tc ; 2. a quantizer; 3. an inverse bit mapper (IBMAP) followed by a parallel-to-serial (P/S) converter. As indicated by the sampling theorem, the choice of the sampling frequency Fc D 1=Tc is related to the bandwidth of the signal s.t/ (see (1.142)). In practice, there is a trade-off between the complexity of the anti-aliasing lter and the choice of the sampling frequency, which must be greater than twice the signal bandwidth. For audio signals, Fc depends on the signal quality that we wish to maintain and therefore it depends on the application, see Table 5.1 [2].
Figure 5.9. Basic scheme for the digital transmission of an analog signal.
Application telephone 300 broadcasting 50 audio, compact disc digital audio tape
Passband (Hz)
Fc (Hz)
3400 (narrow band speech) 8000 7000 (wide band speech) 16000 10 20000 44100 10 20000 48000
338
The choice of the quantizer parameters is somehow more complicated and will be dealt with in detail in the following sections. We will consider now the quantizer as an instantaneous non-linear transformation that maps the real values of s in a nite number of values of sq . To illustrate the principle of an ADC, let us assume that sq assumes values that are taken from a set of 8 elements:2 Q[s.kTc /] D sq .kTc / 2 fQ
4 ; Q 3 ; Q 2 ; Q 1 ; Q1 ; Q2 ; Q3 ; Q4 g
(5.3)
Therefore sq .kTc / may assume only a nite number of values, which can be represented as binary values, for example, using the inverse bit mapper of Table 5.2. It is convenient to consider the sequence of bits that gives the binary representation of fsq g instead of the sequence of values itself. In our example, with a representation using three bits per sample, the bit rate of the system is equal to Rb D 3Fc (bit/s) (5.4)
The inverse process (decoding) takes place at the receiver: the bit-mapper (BMAP) restores the quantized levels, and an interpolator lter yields an estimate of the analog signal.
Values Integer Binary representation sq .kTc / representation c.k/ D .c2 ; c1 ; c0 / Q 4 Q 3 Q 2 Q 1 Q1 Q2 Q3 Q4 0 1 2 3 4 5 6 7 000 001 010 011 100 101 110 111
The notation adopted in (5.3) to dene the set reects the fact that in most cases the set of values assumed by sq is symmetrical around the origin. 3 From the Observation 1.7 on page 71, if s .kT / is WSS, then the interpolated random process s .t/ is WSS q c q 2 2 and E[sq .t/] D E[sq .kTc /], whenever the gain of the interpolator lter is equal to one. As a result the signalto-noise ratio in (5.1) becomes independent of t and can be computed using the samples of the processes, 2 3q D E[s 2 .kTc /]=E[eq .kTc /].
339
g nTc
t T =1/Fc c
Typically, however, the DAC employs a simple holder that holds the input values as illustrated in Figure 5.10. In this case t Tc =2 D wTc .t/ (5.6) g I .t/ D rect Tc and G I . f / D Tc sinc f Fc e
2 f Tc =2
(5.7)
Unless the sampling frequency has been chosen sufciently higher than twice the bandwidth of s.t/, we see that the lter (5.7), besides not attenuating enough the images of sq .kTc /, introduces distortion in the passband of the desired signal.4 A solution to this Q problem consists in introducing, before interpolation, a digital equalizer lter with a frequency response equal to 1= sinc. f Tc / in the passband of s.t/. Figure 5.11 illustrates the solution. A simple digital equalizer lter is given by G comp .z/ D 9 1 C z 16 8
1
1 z 16
(5.8)
g comp T c
wT
~ (t) sq
In many applications, to simplify the analog interpolator lter, the signal before interpolation is oversampled: for example, by digital interpolation of the signal sq .kTc / by at least a factor of 4. Q
340
Figure 5.12. Frequency responses of a three-coefcient equalizer lter gcomp and of the overall lter gI D gcomp wTc .
An alternative solution is represented by an IIR lter with: G comp .z/ D 9=8 1 C 1=8z
1
(5.9)
whose frequency response is given in Figure 5.13. In the following sections, by the term DAC we mean a digital-to-analog converter with the aforementioned variations.
If L D 2b is the number of levels of the quantizer, then the encoder bit rate is equal to Rb D bFc bit/s. Another important parameter of the binary channel is the bit error probability O Pbit D P[b` 6D b` ] (5.11)
If an error occurs, the reconstructed binary representation c.k/ is different from c.k/: Q consequently the reconstructed level is sq .kTc / 6D sq .kTc /. In the case of a speech signal, Q
341
Figure 5.13. Frequency responses of an IIR equalizer lter gcomp and of the overall lter gI D gcomp wTc .
such an event is perceived by the ear as a fastidious impulse disturbance. For speech signals to have an acceptable quality at the receiver it must be Pbit 10 3 .
5.1.2
At the output of an ADC, the PCM encoded samples, after suitable transformations, can be further quantized in order to reduce the bit rate. From [3], we list in Figure 5.14 various coding techniques, which are divided into three groups, that essentially exploit two elements: redundancy of speech, sensibility of the ear as a function of the frequency. Waveform coding. Waveform encoders attempt to reproduce the waveform as closely as possible. This type of coding is applicable to any type of signal; two examples are the PCM and ADPCM schemes. Coding by modeling. In this case coding is not related to signal samples, but to the parameters of the source that generates them. Assuming the voiced/unvoiced speech model, an example of a classical encoder (vocoder) is given in Figure 5.15, where a periodic excitation, or white noise segment, ltered by a suitable lter, yields a synthesized speech segment. A more sophisticated model uses a more articulated multipulse excitation. Frequency-domain coding. In this case coding occurs after signal transformation to a domain different from time, usually frequency: examples are sub-band coding and transform coding.
342
343
Algorithm
Year
Codebook excited LP Multipulse excited LP Vector quantization Time domain harmonic scaling Adaptive transform coding Sub-band coding Residual excited LP Adaptive predictive coding Formant vocoder Cepstral vocoder Channel vocoder Phase vocoder Linear prediction vocoder Adaptive differential PCM Differential PCM Adaptive delta modulation Delta modulation Pulse code modulation
CELP MELP VQ TDHS ATC SBC RELP APC FOR-V CEP-V CHA-V PHA-V LPC-V ADPCM DPCM ADM DM PCM
1984 1982 1980 1979 1977 1976 1975 1968 1971 1969 1967 1966 1966
Various coding techniques are listed in Table 5.3. Table 5.4 illustrates the characteristics of a few systems, putting into evidence that for more sophisticated encoders the implementation complexity expressed in millions of instructions per second (MIPS), as well as the delay introduced by the encoder (latency), can be considerable. The various coding techniques are different in quality and cost of implementation. With respect to the perceived quality, on a scale from poor to excellent, three categories of encoders perform as illustrated in Figure 5.16: obviously a higher implementation complexity is expected for encoders with low bit rate and good quality. We go from a bit rate in the range from 4.4 to 9.6 kbit/s for cellular radio systems, to a bit rate in the range from 16 to 64 kbit/s for transmission over the public network. Generally a coding technique is strictly related to the application and depends on various factors: signal type (for example speech, music, voice-band data, signalling, etc.); maximum tolerable latency; implementation complexity. In particular, speech encoder applications for bit rate in the range 416 kbit/s are: long distance and satellite transmission; digital mobile radio (cellular radio);
344
Coder
Bit rate Computational Latency (kbit/s) complexity (MIPS) (ms) 64 32 16 8 4 2 0.0 0.1 1 10 100 1 0 0 25 35 35 35
Figure 5.16. Audio quality vs. bit rate for three categories of encoders.
modem transmission over the telephone channel (voice mail); speech storage for telephone services and speech encryption; packet networks with integrated speech and data.
5.2
5.2.1
Instantaneous quantization
Parameters of a quantizer
We consider a sample of a discrete-time random process s.kTc /, obtained by sampling the continuous-time process s.t/ with rate Fc . To simplify the notation we choose Tc D 1, unless otherwise stated. With reference to the scheme of Figure 5.17, for a quantizer with L output values we have: input signal s.k/ 2 <; quantized signal sq .k/ 2 Aq D fQ L=2 ; : : : ; Q the alphabet Aq are called output levels;
1 ; Q1 ; : : : ; Q L=2 g;
the L values of
345
Figure 5.17. Quantization and mapping scheme: (a) encoder, (b) decoder.
code word c.k/ 2 f0; 1; : : : ; L 1g, which represents the value of sq .k/. The system with input s.k/ and output c.k/ constitutes a PCM encoder. The quantizer can be described by the function Q : < ! Aq (5.12)
For a given partition of the real axis in the intervals fRi g, i D L=2; : : : ; 1; 1; : : : ; L=2 S L=2 T such that < D i D L=2; i 6D0 Ri , Ri R j D ; for i 6D j, (5.12) implies the following rule Q[s.k/] D sq .k/ D Qi if s.k/ 2 Ri (5.13)
A common choice for the decision intervals Ri is given by: ( Ri D .i ; i C1 ] for i D L=2; : : : ; 1 Ri D .i
1 ; i ]
for i D 1; : : : ; L=2
(5.14)
where L=2 D 1 and L=2 D 1. We note that the decision thresholds fi g are L 1, being L=2 and L=2 assigned. The mapping rule (5.12) is called the quantizer characteristic and is illustrated in Figure 5.18 for L D 8 and 0 D 0. The L values of sq .k/ can be represented by integers c.k/ 2 f0; 1; : : : ; L 1g or by a binary representation with dlog2 Le bits. For the quantizer characteristic of Figure 5.18, a binary representation is adopted that goes from 000 (the minimum level), to 111 (the maximum level); in this example the bit rate of the system is equal to Fb D 3Fc bit/s. Let eq .k/ D sq .k/ s.k/ (5.15) be the quantization error. From the relation sq .k/ D s.k/ C eq .k/ we have that the quantized signal is affected by a certain error eq .k/. We can formulate the problem as that of representing s.k/ with the minimum number of bits b, to minimize the system bit rate, and at the same time constraining the quantization error, so that a certain level of quality of the quantized signal is maintained. Observation 5.1 In this chapter the notation c.k/ is used to indicate both an integer number and its vectorial binary representation (see (5.18)). Furthermore, in the context of vector quantization the elements of the set Aq are called code words.
346
5.2.2
Uniform quantizers
A quantizer with L D 2b equally spaced output levels and decision thresholds is called uniform. For: ( i D L=2 C 1; : : : ; 1 i C1 i D 1 (5.16) i D 1; 2; : : : ; L=2 1 i i 1 D 1 8 > Qi C1 Qi D 1 < Q1 Q 1 D 1 > : Qi Qi 1 D 1 iD L=2; : : : ; 2 (5.17) i D 2; : : : ; L=2
where 1 is the quantization step size. Two types of characteristics are distinguished, midtread and mid-riser, depending on whether the zero output level belongs or not to Aq . Mid-riser characteristic. The quantizer characteristic is given in Figure 5.19 for L D 8: in this case the smallest value, in magnitude, assumed by sq .k/ is 1=2, even for a very small input value s. Let the binary representation of c.k/ be dened according to the following rule: the most signicant bit of the binary representation of c.k/ denotes the sign .1/ of the input value, whereas the remaining bits denote the amplitude. Therefore adopting the binary vector representation c.k/ D [cb
1 .k/; : : : ; c0 .k/]
c j .k/ 2 f0; 1g
(5.18)
347
7 2 5 2 3 2 2 - 4 - 3 - 2 - 100
sq=Q[s]
010
011
001
000
- 2 3 2 5 2 7 2
101
110
111
the relation between sq .k/ and c.k/ is given by sq .k/ D 1.1 2cb
1 .k// b 2 X jD0
c j .k/ 2 j C
1 .1 2
2cb
1 .k//
(5.19)
Mid-tread characteristic. The quantizer characteristic is shown in Figure 5.20. Zero is a value assumed by sq . Let the binary representation of c.k/ be the twos complement representation of the level number. Then we have sq .k/ D 1 c.k/. Note that the characteristic is asymmetric around zero, hence we may use L 1 levels (giving up the minimum output level), or choose an implementation that can be slightly more complicated than in the case of a symmetric characteristic (see page 357).
Quantization error
We will refer to symmetrical quantizers, with mid-riser characteristic. An example with L D 23 D 8 levels is given in Figure 5.21: in this case the decision thresholds are i D i1, i D L=2 C 1; : : : ; 1; 0; 1; : : : ; L=2 1, with, as usual, L=2 D 1 and L=2 D 1. The output values are given by 8 > iC1 1 > i D L=2; : : : ; 1 < 2 (5.20) Qi D > > i 1 1 : i D 1; : : : ; L=2 2 Correspondingly the decision intervals are given by (5.14).
348
sq=Q[s]
3 010 011
5 2
7 2
110
- 2
101
- 3
100
- 4
We note that if sq .k/ D Qi , then the b 1 least signicant bits of c.k/ are given by the binary representation of .jij 1/, and c.k/ assumes amplitude values that go from 0 to L=2 1 D 2b 1 1. If for each value of s we compute the corresponding error eq D Q.s/ s, we obtain the quantization error characteristic of Figure 5.21. We dene the quantizer saturation value as sat D .L=2/
1
C1
(5.21)
that is shifted by 1 with respect to the last nite threshold value. Then we have jeq j and ( eq D Q L=2 s for s > sat Q L=2 s for s < sat (5.23) 1 2 for jsj < sat (5.22)
Consequently, eq may assume large values if jsj > sat . This observation suggests that the real axis be divided into two parts: 1. the region s 2 . 1; sat / [ .sat ; C1/, where eq is called saturation or overload error .esat /;
349
350
2. the region s 2 [ sat ; sat ], where eq is called granular error .egr /; the interval [ sat ; sat ] is also called quantizer range. It is often useful to compactly represent the quantizer characteristic in a single axis, as illustrated in Figure 5.21c, where the values of the decision thresholds are indicated by dashed lines, and the quantizer output values by dots.
Relation between
, b and sat
The quantization step size 1 is chosen so that 2sat D L1 Therefore, for L D 2b , 1D 2sat 2b (5.25) (5.24)
If js.k/j < sat , observing (5.22) this choice guarantees that eq is granular with amplitude in the range 1 1 eq .k/ 2 2 (5.26)
If js.k/j > sat the saturation error can assume large values: therefore sat must be chosen so that the probability of the event js.k/j > sat is small. For a xed number of bits b, and consequently for a xed number of levels L, it is important to verify that, increasing sat , 1 also increases and hence also the granular error; on the other hand, choosing a small 1 leads to a considerable saturation error. As a result, for each value of b there will be an optimum choice of sat and hence of 1. In any case, to decrease both errors we must increase b with consequent increase of the encoder bit rate.
351
p (a) e
q
1 __
__ 2
__ 2
We note that if s.k/ is a constant signal the above assumptions are not true; they hold in practice if fs.k/g is described by a function that signicantly deviates from a constant and 1 is adequately small, that is b is large. Figure 5.24 illustrates the quantization error for a 16-level quantized signal. The signal eq .t/ is quite different from s.t/ and the above assumptions are plausible. If the probability density function of the signal to quantize is known, letting g denote the function that relates s and eq , that is eq D g.s/, also called quantization error characteristic, the probability density function of the noise is obtained as an application of the theory of functions of a random variable, that yields X ps .b/ 1 1 <a< (5.30) peq .a/ D jg 0 .b/j 2 2 1
b2g .a/
where g is the inverse of the error function, or equivalently the set of values of s corresponding to a given value of eq . We note that in this case the slope of the function g is always equal to one, hence g 0 .b/ D 1, and from (5.15) for 1=2 < a < 1=2 we get L L 1 g .a/ D Qi a; i D ; : : : ; 1; 1; : : : ; (5.31) 2 2 Finally, peq .a/ D
iD 2 X L 2 ; i 6D0 L
1 ./
ps .Qi
a/
1 1 <a< 2 2
(5.32)
It can be shown that, if 1 is small enough, the sum in (5.32) gives origin to a uniform function peq , independently of the form of ps .
352
Choosing sat so that eq is granular, from (5.29) we get Meq ' Megr ' 12 12 (5.34)
For an exact computation that includes also the saturation error we need to know the probability density function of s. The statistical power of eq is given by Z C1 2 Meq DE[eq .k/] D [Q.a/ a]2 ps .a/ da Z D
1 sat Zsat1 sat
[Q.a/ [Q.a/
sat
[Q.a/
1
a]2 ps .a/ da
(5.35)
In (5.35) the rst term is the statistical power of the granular error, Megr , and the other two terms express the statistical power of the saturation error, Mesat . Let us assume that ps .a/
353
is even and the characteristic is symmetrical, i.e. i D i and Q i D Qi ; then we get 8 9 L >X Z > Z sat <2 1 i = (5.36) .Qi a/2 ps .a/ da C .Q L a/2 ps .a/ da Megr D 2 2 > i D1 i 1 > L : ;
2
1
Z Mesat D 2
C1
sat
.Q L
2
a/2 ps .a/ da
(5.37)
If the probability of saturation satises the relation P[js.k/j > sat ] 1, then Mesat ' 0; introducing the change of variable b D Qi a, as i D Qi C 1=2 and i 1 D Qi 1=2, we have L=2 X Z 1=2 b2 ps .Qi b/ db (5.38) Meq ' Megr D 2
i D1 1=2
1 2
(5.39)
ps .b/ db D 1, we get !Z
1=2 1=2
L=2 X i D1
ps .Qi /1
12 b2 db ' 1 12
(5.40)
assuming that sat is large enough, so that the saturation error is negligible, and 1 is sufciently small to verify (5.39).
the procedure of designing a uniform quantizer consists of three steps. 1. Determine sat so that the saturation probability is sufciently small: Psat D P[js.k/j > sat ] 1 (5.43)
354
Psat
8 > 0:046 > > > > > < sat D 0:0027 D 2Q > s > > > > > 0:000063 :
2. Choose L so that the signal-to-quantization error ratio assumes a desired value 3q D 3. Given L and k f , we obtain 1D 2s 2sat D L kf L (5.45) Ms 2 ' 2 s D 3k 2 L 2 f Meq 1 =12 (5.44)
(5.46)
Recalling that this law considers only granular error, if we double the number of quantizer levels for a given loading factor, i.e. increase by one the number of bits b, the signal-toquantization error ratio increases by 6 dB. Example 5.2.1 Let s.k/ 2 U[ smax ; smax ]. Setting sat D smax , we get p smax sat D D 3 H) .3q /d B D 6:02 b s s Example 5.2.2 Let s.k/ D smax cos.2 f 0 Tc k C '/. Setting sat D smax , we get p smax sat D D 2 H) .3q /d B D 6:02 b C 1:76 s s
(5.47)
(5.48)
Example 5.2.3 For s.k/ not limited in amplitude, and assuming Psat negligible for sat D 4s , we get .3q /d B D 6:02 b
6
7:2
(5.49)
355
45
40
35
30
b=8
q (dB)
25
20
15
10
b=8
60
50
40
30 s/ sat (dB)
20
10
Figure 5.25. Signal-to-quantization error ratio versus s =sat of a uniform quantizer for granular noise only (dashed lines), for a Laplacian signal (dashed-dotted lines), and of a -law ( D 255) quantizer (continuous lines). The parameter b is the number of bits of the quantizer. The expression of 3q for granular noise only is given by (5.46) and (5.64) for a uniform and a -law quantizer, respectively.
The plot of 3q , given by (5.46), versus the statistical power of the input signal is illustrated in Figure 5.25 for various values of b. We note that for values of s near sat the approximation Meq ' Megr is no longer valid because Mesat becomes non-negligible. For the computation of Mesat we need to know the probability density function of s and apply (5.35). Assuming a Laplacian signal we obtain the curves also shown in Figure 5.25, that coincide with the curves given by (5.46) for s sat . The optimization of 3q for the uniform quantization of a signal with a specied amplitude distribution yields the results given in Table 5.5 [4]. We note that for the more dispersive inputs the optimum value of 1 increases, and consequently the value of 3q decreases. We also note that the quantizers obtained by the optimization procedure and by the method on page 353 are in general different. Example 5.2.4 For s.k/ 2 N 0; s2 and b D 5, observing Table 5.5 we have optimum performance for 1=s D 0:1881, and consequently sat D 2b 1 1 D 3:05s . As shown in Figure 5.26, the optimum value of 3q is obtained by determining the minimum of .Megr C Mesat /=Ms as a function of s =sat . The optimum point depends on b: we have sat D `s , where ` increases with b; in particular for b D 3 it turns out sat D 2:3 s , whereas for b D 8 we obtain sat D 3:94 s .
356
Table 5.5 Optimal quantization step size and maximum corresponding value 3q of a uniform quantizer for different ps .a/ (U: uniform, G: Gaussian, L: Laplacian, : gamma). [From Jayant and Noll (1984).]
b (bit/sample) U 1 2 3 4 5 6 7 8
3
1opt =s ps .a/ G 1.5956 0.9957 0.5860 0.3352 0.1881 0.1041 0.0569 0.0308 L 1.4142 1.0874 0.7309 0.4610 0.2800 0.1657 0.0961 0.0549 1.1547 1.0660 0.7957 0.5400 0.3459 0.2130 0.1273 0.0743 U 6.02 12.04 18.06 24.08 30.10 36.12 42.14 48.17
max.3q /d B ps .a/ G 4.40 9.25 14.27 19.38 24.57 29.83 35.13 40.34 L 3.01 7.07 11.44 15.96 20.60 25.36 30.23 35.14 1.76 4.95 8.78 13.00 17.49 22.16 26.99 31.89
x 10
eq
/M
3
Me
gr
/Ms
Me
sat
/Ms
0 0.25
0.3
s/sat
0.35
0.4
2 Figure 5.26. Determination of the optimum value of 3q for b D 5 and s.k/ 2 N .0; s /.
We conclude this section observing that for a non-stationary signal, for example, a voice signal, setting sat D 4s , where s2 is computed for a voiced spurt, yields 3q ' 33 dB for b D 7, good enough for telephone communications. However, in an unvoiced spurt s2 can be reduced by 2030 dB, and consequently 3q is degraded by an amount equivalent to 35 bit.
357
358
clock
stop stop
1.
2. A second possible implementation, which encodes one bit at a time, is given in Figure 5.29. In this case b 1 comparisons are made: it is as if we were to explore a complete binary tree whose 2b 1 leaves represent the output levels. For example, for b D 3, neglecting the sign bit, the code word length is 2, and c.k/ D .c1 ; c0 /. To determine the bits c0 and c1 we can operate as follows: if V < 2 ) c1 D 0 if V < 1 C c1 21 1 ) c0 D 0 otherwise c1 D 1 otherwise c0 D 1
Only two comparisons are made, but the decision thresholds now depend on the choice of the previous bits. 3. The last implementation, which encodes one code word of (b 1) bit at a time, is given in Figure 5.30. In this scheme V is compared simultaneously with the 2b 1 quantizer thresholds: the outcome of this comparison is a word of 2b 1 bit formed by a sequence of 0 followed by a sequence of 1; through a logic network this word is mapped to a binary word of b 1 bits that yields the PCM encoding of s.k/. These encoders are called ash converters. We conclude this section explaining that the acronym PCM stands for pulse code modulation. We waited until the end of the section to avoid confusion about the term modulation: in fact, PCM is not a modulation, but rather a coding method.
5.3
Non-uniform quantizers
There are two observations that suggest the choice of a non-uniform quantizer. The rst refers to stationary signals with a non-uniform probability density function: for such signals
359
1 s(k)=V 2 2 3
b-1
2b-1
uniform quantizers are suboptimum. The second refers to non-stationary signals, e.g., speech, for which the ratio between instantaneous power (estimated over windows of tenths of milliseconds) and average power (estimated over the whole signal) can exhibit variations of several dB; moreover, the variation of the average power over different links is also of the order of 40 dB. Under these conditions a quantizer with nonuniform characteristics, as that depicted for example in Figure 5.31, is more effective because the signal-to-quantization error ratio 3q is almost independent of the instantaneous power. As also illustrated in Figure 5.31, for a non-uniform quantizer the quantization error is large if the signal is large, whereas it is small if the signal is small: as a result the ratio 3q tends to remain constant for a wide dynamic range of the input signal.
360
5.3.1
Companding techniques
Figure 5.32b illustrates in detail the principle of Figure 5.32a. The signal is rst compressed through a non-linear function F, that yields the signal y D F.s/ (5.50) In Figure 5.32 we assume sat D 1. If sat 6D 1 we need to normalize s to sat . The signal y is uniformly quantized and the code word given by the inverse bit mapper is transmitted. At the receiver the bit mapper gives yq , that must be expanded to yield a quantized version of s sq D F
1
[Q[y]]
(5.51)
This quantization technique takes the name of companding from the steps of compressing and expanding. We nd that the ideal characteristics of F[] should be logarithmic, F[s] D ln s We consider the two blocks shown in Figure 5.34. (5.52)
361
Figure 5.32. (a) Use of a compression function F to implement a non-uniform quantizer; (b) non-uniform quantizer characteristic implemented by companding and uniform quantization. Here sat D 1 is assumed.
Encoding. that is
362
Figure 5.33. Non-uniform quantizer implemented digitally using a uniform quantizer with small step size followed by a look-up table.
Figure 5.34. Non-uniform quantization by companding and uniform quantization: (a) PCM encoder, (b) decoder.
and assume the sign of the quantized signal is equal to that of s.k/. The quantization of y.k/ yields yq .k/ D Q[y.k/] D ln js.k/j C eq .k/ The value c.k/ is given by the inverse bit mapping of yq .k/ and the sign of s.k/. Decoder. Assuming c.k/ is correctly received, observing (5.55), the quantized version of s.k/ is given by sq .k/ D e yq .k/ sgn[s.k/] D js.k/j sgn[s.k/]eeq .k/ D s.k/eeq .k/ If eq 1, then eeq .k/ ' 1 C eq .k/ (5.57) (5.56) (5.55)
363
where eq .k/s.k/ represents the output error of the system. As eq .k/ is uncorrelated with the signal ln js.k/j, and hence with s.k/ (see (5.28)), we get 3q D Ms 2 .k/s 2 .k/] E[eq D 1 1 D 2 .k/] Meq E[eq (5.59)
where from (5.41) we have that Meq depends only on the quantization step size 1. Consequently 3q does not depend on Ms . We note that a logarithmic compression function generates a signal y with unbounded amplitude, thus an approximation of the logarithmic law is usually adopted. Regulatory bodies have dened two compression functions: 1. A-law (A D 87:56). For sat D 1, 8 Ajsj 1 > > 0 jsj < 1 C ln.A/ A (5.60) y D F[s] D > 1 C ln.Ajsj/ 1 > : jsj 1 1 C ln.A/ A This law, illustrated in Figure 5.35 for two values of A, is adopted in Europe. The sign is considered separately: sgn[y] D sgn[s]
1
(5.61)
0.9
0.8
0.7
A=87.56
0.6
F(s)
0.3
0.2
0.1
0.1
0.2
0.3
0.4
0.5 s
0.6
0.7
0.8
0.9
364
0.6
F(s)
=5
0.5
0.4 =0 0.3
0.2
0.1
0.1
0.2
0.3
0.4
0.5 s
0.6
0.7
0.8
0.9
2. -law ( D 255). For sat D 1, ln.1 C jsj/ (5.62) ln.1 C / This law, illustrated in Figure 5.36 for four values of , is adopted in the United States and Canada. The compression increases for higher values of ; the standard value of is equal to 255. We note that, for s 1, we have ln[ s] F[s] D (5.63) ln.1 C / as in the ideal case (5.54). Similar behavior is exhibited by (5.60). y D F[s] D
365
uniform quantizer with Laplacian input. We emphasize that also in this case 3q increases by 6 dB with the increase of b by one. We also note that, if b D 8, 3q ' 38 dB for a wide range of values of s . An effect not shown in Figure 5.25 is that, by increasing , the plot of 3q becomes atter, but the maximum value decreases. Observation 5.2 In the standard non-linear PCM, a quantizer with 128 levels (7 bit/sample) is employed after the compression; including also the sign we have 8 bit/sample. For a sampling frequency of Fc D 8 kHz, this leads to a bit rate of the system equal to Rb D 64 kbit/s.
Digital compression
An alternative method to the compression-quantization scheme is illustrated by an example in Figure 5.37. The relation between s.k/ and yq .k/ is obtained through a rst multi-bit (5 in gure) quantization to generate xq ; then we have a mapping of the 5 bits of xq to the 3 bits of yq using the mapper (sign omitted) of Table 5.6. For decoding, for each code word yq we select only one code word xq , which represents the reconstructed value sq . Using the standard compression laws, we need to approximate the compression functions by piecewise linear functions, as shown in Figure 5.39. For encoding, a mapper with 12-bit input and 8-bit output is given in Table 5.7. For decoding, we select for each compressed
Figure 5.37. Distribution of quantization levels for a 3-bit -law quantizer with D 40.
366
Coding of xq 0000 0001 0010 0011 0100 0101 1000 1001 1010 1011 1100 1101 1110 1111
Coding of yq 00 01
10
0100
11
1011
code word a corresponding linear code word, as given in the third column of Table 5.7. In the literature there are other non-linear PCM tables, that differ in the compression law or in the accuracy of the codes [4].
5.3.2
Assuming we know the probability density function of the input signal s.k/, stationary with variance s2 , we desire to determine the parameters of the non-uniform quantizer that optimizes 3q . The problem, illustrated in Figure 5.40, consists in choosing the decision thresholds ; : : : ; ; ; ; : : : ; L (5.65) L D 1 L D C1 L 1 0 1
2 1 2 1 2 2
and the quantization levels L L ; : : : ; 1; 1; : : : ; (5.66) 2 2 that minimize the statistical power of the error (minimum mean-square error criterion) fQi g iD Meq D E[.sq .k/ s.k//2 ] D E[.Q[s.k/] s.k//2 ] (5.67)
367
Figure 5.38. Piecewise linear approximation of the A-law compression function (A D 87:6). The 12-bit encoded input signals are mapped into 8-bit signals. Table 5.7 Non linear PCM from 11 to 7 bits (sign omitted).
Compressed code .yq / 111WXYZ 110WXYZ 101WXYZ 100WXYZ 011WXYZ 010WXYZ 001WXYZ 000WXYZ
Assuming ps .a/ even, because of the symmetry of the problem we can halve the number of variables to be determined by setting ( L i D i 1 i D 1; : : : ; (5.68) 2 0 D 0 L (5.69) i D 1; : : : ; Q i D Qi 2 and L=2 X Z i Meq D 2 .Qi a/2 ps .a/ da (5.70)
i D1 i
1
368
Necessary but not sufcient conditions for minimizing (5.67) are @Meq @i @Meq @Qi D0 D0 i D 1; : : : ; i D 1; : : : ; L 2 L 2 1 (5.71) (5.72)
369
p (a) s
-4 =-
-3 -2 -1 0
3 Q4
4=+
Q-4
Figure 5.40. Decision thresholds and output levels for a particular ps .a/ (b D 3).
i /2 ps .i /
.Qi C1
i /2 ps .i /
(5.73)
2Qi i
2 Qi C1
i2 C 2Qi C1 i ] D 0
(5.74)
Qi C Qi C1 2
(5.75)
i
1
.Qi
a/ ps .a/ da D 0
(5.76)
In other words, (5.75) establishes that the optimal threshold lies in the middle of the interval between two adjacent output values, and (5.77) sets Qi as the centroid of ps ./ in the interval [i 1 ; i ]. These two rules are illustrated in Figure 5.41.
Max algorithm
We present now the Max algorithm to determine the decision thresholds and the optimum quantization levels. 1. Fixed Q1 at random, we use (5.77) to get 1 from the integral equation Z 1 aps .a/ da 0 Q1 D Z 1 ps .a/ da
0
(5.78)
370
ps(a)
i-1 Qi
i Qi+1
i+1
Figure 5.41. Optimum decision thresholds and output levels for a given ps .a/.
2. From (5.75) we obtain Qi C1 D 2i C Qi for i D 1. 3. We use (5.77) to obtain i C1 by the equation Z iC1 aps .a/ da i Qi C1 D Z iC1 ps .a/ da
i
(5.79)
2. For i D .L=2/
2
1 we obtain (5.80)
C QL
L 1 Z2 C1 L 2
(5.81) ps .a/ da
then the parameters determined are optimum. Otherwise, if (5.81) is not satised we must change our choice of Q1 in step 1) and repeat the procedure.
Lloyd algorithm
This algorithm uses (5.75) and (5.77), but in a different order. 1. We set a relative error > 0 and D0 D 1. 2. We choose an initial partition of the positive real axis: P1 D f0 ; 1 ; : : : ; L=2 D C1g such that 0 D 0 < 1 < < L=2 D C1. (5.82)
371
3. We set the iteration index j at 1. 4. We obtain the optimum alphabet A j D fQ1 ; : : : ; Q L=2 g for the partition P j using (5.77). 5. We evaluate the distortion associated with the choice of P j and A j ,
2 D j D E[eq ] D 2 L=2 XZ i D1 i
1
.Qi
a/2 ps .a/ da
(5.83)
6. If Dj
1
Dj
Dj
< j C 1.
(5.84)
7. We derive the optimum partition P j D f0 ; 1 ; : : : ; L=2 D C1g for the alphabet A j 1 using (5.75). 8. We go back to step 4. We observe that the sequence D j > 0 is non-increasing: hence the algorithm is converging, however, not necessarily to the absolute minimum, unless some assumptions are made about ps ./.
Expression of
For both algorithms it is important to initialize the various parameters near the optimum values. The considerations that follow have this objective, in addition to determining the optimum value of 3q for a non-uniform quantizer, at least approximately for a number of bits sufciently high. From (5.70), assuming that ps .a/ ' ps .i we have that Meq D 2 '2
L=2 XZ i D1 L=2 X i D1 i
1
1/
for i
a < i
(5.85)
.Qi
1/
a/2 ps .a/ da Z
i
1
(5.86) .Qi a/ da
2
ps .i
372
and Z Qi D Z
i i
a da
1 i
D da
1
i C i 2
(5.88)
i D 1; : : : ;
L 2
(5.89)
Meq D 2
ps .i
1/
.1i /3 12
(5.90)
It is now a matter of nding the minimum of (5.90) with respect to .1i /, with the constraint that the decision intervals cover the whole positive axis; this is obtained by imposing that
L=2 X i D1
ps .i
1=3
1 / .1i / ' 2
Z
0
C1
ps .a/ da D K
1=3
(5.91)
min
Meq C K
L=2 X i D1
!#
1=3 ps .i 1 / .1i /
(5.92)
with Meq given by (5.90). By setting to zero the partial derivative of (5.92) with respect to .1i /, we obtain ps .i that yields p .1i / D 2 ps Substituting (5.94) in (5.91) yields p K D 2L (5.95)
1=3 1/
.1i /2 1=3 C . ps .i 4
1 //
D0
i D 1; : : : ;
L 2
(5.93)
.i
1/
(5.94)
373
hence .1i / D K ps L
1=3
.i
1/
i D 1; : : : ;
L 2
(5.96)
For a quantizer optimized for a certain probability density function, and for a high number of levels L D 2b (so that (5.85) holds), we have 3q D Ms Meq;opt D 22b ff (5.98)
where f f is a form factor related to the amplitude distribution of the normalized signal s .k/ D s.k/=s , Q Q K3 ff D 12 Q K D Z
C1 1
ps .a/ da Q
1=3
(5.99)
p Q In the Gaussian case, s.k/ 2 N .0; s2 /, s .k/ 2 N .0; 1/, and f f D 2=. 3 /. Actually (5.96) indicates that the optimal thresholds are concentrated around the peak of the probability density; moreover, the optimum value of 3q , according to (5.98), follows the increment law of 6 dB per bit, as in the case of a quantizer granular error. Observation 5.3 Equation (5.90) can be used to evaluate approximately Meq for a general quantizer characteristic, even of the A-law and -law types. In this case, from Figure 5.32, the quantization step size 1 of the uniform quantizer is related to the compression law according to the relation 1 D F.i / F.i
1/
' .1i / F 0 .i
1/
(5.100)
where F 0 is the derivative of F. Obviously (5.100) assumes that F 0 does not vary considerably in the interval .i 1 ; i ]. Substituting (5.100) in (5.90) we have Meq D 2
L=2 X i D1
ps .i
1/
1 F 0 .i 1 /
.1i / 12
(5.101)
For L sufciently large, the intervals become small and we get Meq ' 12 2 12 Z
0 sat
ps .a/ da [F 0 .a/]2
(5.102)
374
where 1 is related to a uniform quantizer parameters according to (5.25). It is left to the reader to show that for a uniform signal s in [ sat ; sat ], quantized according to the -law, the ratio 3q D Ms =Meq has the expression given in (5.64).
for both uniform and non-uniform quantizers [4]. The optimum value of 3q follows the 6b law only in the case of non-uniform quantizers for b 4. In the case of uniform quantizers, with increasing b the maximum of 3q occurs for a smaller ratio s =sat (due to the saturation error): this makes 13q vary with b and in fact it increases. Finally, we consider what happens if a quantizer, optimized for a specic input distribution, has a different type of input. For example, a uniform quantizer, best
Table 5.8 Optimum quantizers for a signal with Gaussian dis2 tribution (ms D 0, s D 1). [From Jayant and Noll (1984).]
L 2 4 8 16
i i Qi i Qi i Qi i Qi 1 1 0.798 0.453 0.982 0.501 0.245 0.258 0.128 2 1 1.510 1.050 0.756 0.522 0.388 3 1.748 1.344 0.800 0.657 4 1 2.152 1.099 0.942 5 1.437 1.256 6 1.844 1.618 7 2.401 2.069 8 1 2.733 Meq 0.363 0.117 0.0345 0.00955 3q (dB) 4.40 9.30 14.62 20.20
375
Table 5.9 Optimum quantizers for a signal with Laplacian dis2 tribution (ms D 0, s D 1). [From Jayant and Noll (1984).]
L 2 4 8 16
i i Qi i Qi i Qi i Qi 1 1 0.707 1.127 0.420 0.533 0.233 0.264 0.124 2 1 1.834 1.253 0.833 0.567 0.405 3 2.380 1.673 0.920 0.729 4 1 3.087 1.345 1.111 5 1.878 1.578 6 2.597 2.178 7 3.725 3.017 8 1 4.432 Meq 0.500 0.1761 0.0545 0.0154 3q (dB) 3.01 7.54 12.64 18.12
Table 5.10 Optimum quantizers for a signal with gamma distri2 bution (ms D 0, s D 1). [From Jayant and Noll (1984).]
L 2 4 8 16
i i Qi i Qi i Qi i Qi 1 1 0.577 1.268 0.313 0.527 0.155 0.230 0.073 2 1 2.223 1.478 0.899 0.591 0.387 3 3.089 2.057 1.051 0.795 4 1 4.121 1.633 1.307 5 2.390 1.959 6 3.422 2.822 7 5.128 4.061 8 1 6.195 Meq 0.6680 0.2326 0.0712 0.0196 3q (dB) 1.77 6.33 11.47 17.07
for a uniform input, will have very low performance for an input signal with a very dispersive distribution; on the contrary, a non-uniform quantizer, optimized for a specic distribution, can have even higher performance for a less dispersive input signal.
376
16
14
12
10
q (dB)
L
8
G
4
U
0
1
4 b
Figure 5.42. Performance comparison of uniform (dashed line) and non-uniform (continuous line) quantizers, optimized for a specic probability density function of the input signal. Input type: uniform (U), Laplacian (L), Gaussian (G) and gamma (0) [4]. [From Jayant and Noll (1984).]
Figure 5.43. Comparison of the signal-to-quantization error ratio for uniform quantizer (dashed-dotted line), -law (continuous line) and optimum non-uniform quantizer (dotted line), for Laplacian input. All quantizers have 32 levels (b D 5) and are optimized for s D 1.
377
The 0 quantizers have performance that is almost independent of the type of input. The performance also does not change for a wide range of the signal variance, as their characteristic is of logarithmic type (see Section 5.3.1). A comparison between uniform and non-uniform quantizers with Laplacian input is given in Figure 5.43. All quantizers have 32 levels (b D 5) and are determined using: a) Table 5.5 for the uniform Laplacian type quantizer; b) Table 5.9 for the non-uniform Laplacian type quantizer; c) the ( D 255) compression law of Figure 5.36 with sat =s D 1. We note that the optimum non-uniform quantizer gives best performance, even if this happens in a short range of values s ; for a decrease in the input statistical power, performance decreases according to the law 10 log Ms D 20 log s (dB), as we can see from (5.98). Only a logarithmic quantizer is independent of the input signal level.
5.4
Adaptive quantization
An alternative method to quantize a non-stationary signal consists in using an adaptive quantizer. The corresponding coding scheme, which has parameters that are adapted (over short periods) to the level of the input signal, is called an adaptive PCM or APCM.
General scheme
The overall scheme is given in Figure 5.44, where c.k/ 6D c.k/ if errors are introduced Q by the binary channel. For a uniform quantizer, the idea is that of varying with time the quantization step size 1.k/ so that the quantizer characteristic adapts to the statistical power of the input signal. If 1.k/ is the quantization step size at instant k, with reference to Figure 5.21 the quantizer characteristic is dened as
378
Chapter 5. Digital representation of waveforms 8 > i C 1 1.k/ > < 2 output levels: Qi .k/ D > > i 1 1.k/ : 2 thresholds: i .k/ D i 1.k/ iD
L L C 1; : : : ; 1; 0; 1; : : : ; 2 2
L ;:::; 1 2 L i D 1; : : : ; 2 iD
(5.104) 1
If 1opt is the optimum value of 1 for a given amplitude distribution of the input signal assuming s D 1 (see Table 5.5), and s .k/ is the standard deviation of the signal at instant k, then we can use the following rule 1.k/ D 1opt s .k/ (5.105)
For a non-uniform quantizer, we need to change the levels and thresholds according to the relations: Qi .k/ D Qi;opt s .k/ i .k/ D i;opt s .k/ (5.106)
where fQi;opt g and fi;opt g are given in Tables 5.8, 5.9, and 5.10 for various input amplitude distributions. As illustrated in Figure 5.45, an alternative to the scheme of Figure 5.44 is the following: the quantizer is xed and the input is scaled by an adaptive gain g, so that a signal fy.k/g 2 is generated with a constant statistical power, for example, y D 1. Therefore we let g.k/ D 1 s .k/ (5.107)
However, both methods require computing the statistical power s2 of the input signal. The adaptive quantizers are classied as:
379
feedforward, if s is estimated by observing the signal fs.k/g itself; feedback, if s is estimated by observing fsq .k/ D Q[s.k/]g or fc.k/g, i.e. the signals at the output of the quantizer.
5.4.1
The feedforward methods for the two adaptive schemes of Figure 5.44 and Figure 5.45 are shown, respectively, in Figure 5.46 and Figure 5.47. The main difculty in the two methods is that we need to quantize also the value of s .k/ so that it can be coded and transmitted over a binary channel. We emphasize that: 1. because of digital channel errors on both c.k/ and .s .k//q (or gq .k/) it may happen that sq .k/ 6D sq .k/; Q 2. we need to determine the update frequency of s .k/, that is what frequency is required to sample s , and how many bits must be used to represent .s .k//q ; 3. the system bit rate is now the sum of the bit rate of c.k/ and .s .k//q (or gq .k/).
Figure 5.46. APCM scheme with feedforward adaptive quantizer: a) encoder, b) decoder.
(a)
Figure 5.47. APCM scheme with feedforward adaptive gain and xed quantizer: a) encoder, b) decoder.
380
The data sequence that represents f.s .k//q g or fgq .k/g is called side information. Two methods to estimate s2 .k/ are given in Section 1.11.1. For example, using a rectangular window of K samples, from (1.462) we have s2 .k D/ D 1 K
k X nDk .K 1/
s 2 .n/
(5.108)
where D expresses a certain lead of the estimate with respect to the last available sample: typically D D .K 1/=2 or D D K 1. If D D K 1, K samples need to be stored in a buffer and then the average power must be computed: obviously, this introduces a latency in the coding system that is not always tolerable. Moreover, windows usually do not overlap, hence s is updated every K samples. For an exponential lter instead, from (1.468) we have s2 .k D/ D as2 .k 1 D/ C .1 a/s 2 .k/ (5.109)
Typically in this case we choose D D 0. To determine the update frequency of s2 .k/, we .1 a/ recall that the 3 dB bandwidth of s2 .k/ in (5.109) is equal to B D .2 Tc / , for a > 0:9. Typically, however, we prefer to determine a from the equivalence (1.471) with the length of the rectangular window, that gives a D 1 K 1 1 : this means decimating, quantizing, and coding the values given by (5.109) every K instants. In Table 5.11 we give, for three values of a, the corresponding values of K 1 and B for 1=Tc D 8 kHz.
Performance
With the constraint that s varies within a specic range, min s max , in order to keep 3q relatively constant for a change of 40 dB in the input level, it must be max 100 min (5.110)
Actually min controls the quantization error level for small input values (idle noise), whereas max controls the saturation error level. For speech signals sampled at 8 kHz, Table 5.12 shows the performance of different xed and adaptive 8-level (b D 3) quantizers. The estimate of the signal power is obtained by a rectangular window with D D K 1; the decimation and quantization of s2 .k/
Table 5.11 Time constant and bandwidth of a discrete-time exponential lter with parameter a and sampling frequency 8 kHz.
a K 1 1 1 2 2 2
5 6 7
381
Table 5.12 Performance comparison of xed and adaptive quantizers for speech.
3q (dB) Non-adaptive Adaptive K D 128 (16 ms) Adaptive K D 1024 (128 ms)
non-uniform Q law ( D 100, sat =s D 8) Gaussian (3q;opt D 14:6 dB) Laplacian (3q;opt D 12:6 dB) uniform Q Gaussian (3q;opt D 14:3 dB) Laplacian (3q;opt D 11:4 dB) 9.5 7.3 9.9 6.7 7.4 15 13.3 14.7 13.4 12.1 12.8 11.3 11.5
are not considered. Although b D 3 is a small value to draw conclusions, we note that using an adaptive Gaussian quantizer with K D 128 we get 8 dB improvement over a non-adaptive quantizer. If K 128 the side information becomes excessive, conversely there is a performance loss of 3 dB for K D 1024.
5.4.2
As illustrated in Figure 5.48, the feedback method estimates s from the knowledge of fsq .k/ D Q[s.k/]g or fc.k/g. We make the following observations: there is no need to transmit s .k/; therefore feedback methods do not require the transmission of side information; a transmission error on c.k/ affects not only the identication of the quantized level, but also the scaling factor s .k/. Concerning the estimate of s , a possible method consists in applying (5.108) or (5.109), where fs.n/g is substituted by fsq .n/g. However, this signal is available only for n k 1:
382
Pk 1 2 this implies that the estimate (5.108) becomes s2 .k/ D 1=K nD.k 1/ .K 1/ sq .n/, with q a lag of one sample. Likewise, the recursive estimate (5.109) becomes s2 .k/ D as2 .k q q 2 1/ C .1 a/sq .k 1/. Because of the lag in estimating the level of the input signal and the computational complexity of the method itself, we present now an alternative method to estimate s adaptively.
Estimate of s (k)
For an input with s D 1 we compute the discrete amplitude distribution of the code words for a quantizer with 2b levels and jc.k/j 2 f1; 2; : : : ; L=2g. As illustrated in Figure 5.49 for b D 3, let Z opt;1 8 > P[jc.k/j D 1] D 2 > ps .a/ da D pc1 > > > 0 D0 < : : : : (5.111) : : > Z C1 > > > > ps .a/ da D pc4 : P[jc.k/j D 4] D 2
opt;3
If s changes suddenly, the distribution of jc.k/j will be very different with respect to (5.111). For example, if s < 1 it will be P[jc.k/j D 1] pc1 , while P[jc.k/j D 4] pc4 .
0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0
ps (a)
s =1
Q1,opt Q2,opt
Q3,opt
Q4,opt
Figure 5.49. Output levels and optimum decision thresholds for Gaussian s.k/ with unit variance (b D 3).
383
Figure 5.50. Adaptive quantizer where s is estimated using the code words.
The objective is therefore that of changing sq .k/ so that the optimal distribution is obtained for jc.k/j. The algorithm proposed by Jayant [4], illustrated in Figure 5.50, is given by sq .k/ D p[jc.k 1/j]sq .k 1/ (5.112)
in which fp[i]g, i D 1; : : : ; L=2, are suitable parameters. For example, if it is jc.k 1/j D 1, then sq must decrease to reduce the quantizer range, thus p[1] < 1; if instead it is jc.k 1/j D L=2, then sq must increase to extend the quantizer range, thus p[L=2] > 1. In practice, what we do is vary sq by small steps imposing bounds to the variations, that is: min sq .k/ max (5.113)
The problem consists now in choosing the parameters fp[i]g, i D 1; : : : ; L=2. Intuitively it should be .p[1]/ pc1 .p[2]/ pc2 : : : .p[L=2]/ In fact, from (5.112) it follows that ln sq .k/ D ln p[jc.k from which E[ln sq ].k/ D E[ln p[jc.k 1/j]] C E[ln sq .k 1/] (5.116) 1/j] C ln sq .k 1/ (5.115)
pc L=2
D1
(5.114)
In steady state we expect that E[ln sq .k/] D E[ln sq .k E[ln p[jc.k as in (5.114). 1/j]] D
L=2 X i D1
pci ln p[i] D 0
(5.117)
384
Based on numerous tests on speech signals, Jayant also gave the values of the parameters fp[i]g, i D 1; : : : ; L=2. Let 2i q.i/ D L 1 2 1 L 1 ; 1 L 3 1 ;:::;1 (5.118)
In Figure 5.51 the values of fp[i]g are given in correspondence of fq.i/g, i D 1; : : : ; L=2 [4]. For example, for L D 8 the values of fp[i]g, i D 1; : : : ; 4, are in correspondence of the values of fq.i/g D f1=7; 3=7; 5=7; 1g. Therefore p[1] is in the range from 0.8 to 0.9, and p[4] in the range from 1.8 to 2.9; we note that there is a large interval of possible values for p[i], especially if the index i is large. Summarizing, at instant k, sq .k/ is known, and by (5.106) the decision thresholds fi .k/g are also known. From the input sample s.k/, c.k/ is produced by the quantizer characteristic (see Figure 5.52). Then sq .k C 1/ is computed by (5.112) and the thresholds are updated: the quantizer is now ready for the next sample s.k C 1/. At the receiver, the possible output values are also known from the knowledge of sq .k/ (see (5.106)). At the reception of c.k/ the index i in (5.106) is determined and, consequently, sq .k/; in turn the receiver updates the value of sq .k C 1/ by (5.112). Experimental measurements on speech signals indicate that this feedback adaptive scheme offers performance similar to that of a feedforward scheme. An advantage of the algorithm of Jayant is that it is sequential, thus it can adapt very quickly to changes in the mean signal level; on the other hand, it is strongly affected by the errors introduced by the binary channel.
p
3
0 0 1
Figure 5.51. Interval of the multiplier parameters in the quantization of the speech signal as a function of the parameters fq.i/g [4]. [From Jayant and Noll (1984).]
385
Q (k)
4
sq=Q[s]
010
011
p[4]
Q (k)
3
p[3]
001
Q (k)
2
p[2]
000
Q (k)
1
p[1] - Q (k)
1
- 3(k)
- 2(k)
- 1(k)
100
1(k)
2(k)
3(k)
p[1]
101
p[2]
110
- Q (k)
2
p[3]
111
- Q (k)
3
p[4]
- Q (k)
4
Figure 5.52. Input--output characteristic of a 3-bit adaptive quantizer. For each output level the PCM code word and the corresponding value of p are given.
5.5
The basic idea consists in quantizing the prediction error signal rather than the signal O itself.7 With reference to Figure 5.53, for a linear predictor with N coefcients, let s .k/ be the prediction signal :8 s .k/ D O
N X i D1
ci s.k
i/
(5.119)
From (2.81) the prediction error is dened as f .k/ D s.k/ Considering the z-transform, let C.z/ D
N X i D1
s .k/ O
(5.120)
ci z
(5.121)
In the following sections, as well as in some schemes of the previous section on adaptive quantization, when processing of the input samples fs.k/g is involved, it is desirable to perform the various operations in the digital domain on a linear PCM binary representation of the various samples, obtained by an ADC. Obviously, the nite number of bits of this preliminary quantization should not affect further processing. To avoid introducing a new signal, the preliminary conversion by an ADC is omitted in all our schemes. 8 The considerations presented in this section are valid for any predictor, even non-linear predictors.
386
s(k)
+ -
f(k)
^ s(k)
s(k) +
-
f(k)
f(k) sq(k) +
f(k) +
+
sq(k)
^ s(k)
c
+
^ s(k)
(a)
(b)
Figure 5.54. (a) Prediction error lter; (b) Inverse prediction error lter.
Recalling (2.81), [1 C.z/] is the prediction error lter. In the case C.z/ D z 1 , that is for a predictor with a single coefcient equal to one, f .k/ coincides with the difference between two consecutive input samples. It is interesting to re-arrange the scheme of Figure 5.53 in the equivalent scheme of Figure 5.54, where sq .k/ is called the reconstruction signal and is given by sq .k/ D s .k/ C f .k/ O (5.124)
Figure 5.54a illustrates how the prediction error is obtained starting from input s.k/ and prediction s .k/, Figure 5.54b shows how to obtain the reconstruction signal from f .k/ and O s .k/, according to (5.124). From (5.120) and (5.124), it is easy to prove that in the scheme O of Figure 5.54 we have sq .k/ D s.k/ that is the reconstructed signal coincides with the input signal. We will now quantize the signal f f .k/g. (5.125)
5.5.1
With reference to the scheme of Figure 5.55, the following relations hold.
387
Figure 5.55. DPCM scheme with quantizer inserted in the feedback loop: (a) encoder, (b) decoder.
ci sq .k C 1
(5.130) i/ (5.131)
ci sq .k C 1
In other words, the quantized prediction error is transmitted over the binary channel. Let eq; f .k/ D f q .k/ f .k/ (5.132) be the quantization error and 3q; f the signal-to-quantization error ratio 3q; f D E[ f 2 .k/] 2 E[eq; f .k/] (5.133)
388
Recalling (5.98) we know that for an optimum quantizer, with the normalization by the standard deviation of f f .k/g, 3q; f is only a function of the number of bits and of the probability density function of f f .k/g. From (5.128) and (5.132), using (5.126), we have sq .k/ D s .k/ C f .k/ C eq; f .k/ D s.k/ C eq; f .k/ O (5.134)
To summarize, the reconstruction signal is different from the input signal, sq .k/ 6D s.k/, and the reconstruction error (or noise) depends on the quantization of f .k/, not of s.k/. Consequently, if M f < Ms then also Meq; f < Meq and the DPCM scheme presents an advantage over PCM. Observing (5.134) the signal-to-noise ratio is given by 3q D Given Gp D called prediction gain, it follows that 3q D G p 3q; f (5.137) Ms Mf (5.136) Ms Ms M f D Meq; f M f Meq; f (5.135)
where observing (5.133) 3q; f depends on the number of quantizer levels, which in turn determine the transmission bit rate, whereas G p depends on the predictor complexity and on the correlation sequence of the input fs.k/g. We observe that the input to the lter that yields s .k/ in (5.129) is fsq .k/g and not fs.k/g; this will cause a deterioration of G p with O respect to the ideal case fsq .k/ D s.k/g. This decrease will be more prominent the larger feq; f g will be with respect to fs.k/g. If we ignore the dependence of G p on feq; f .k/g, (5.137) shows that to obtain a given 3q we can use a quantizer with a few levels, provided the input fs.k/g is highly predictable. Therefore G p can be sufciently high also for a predictor with a reduced complexity. For the quantizer, assuming the distribution of f f .k/g is known, 3q; f is maximized by selecting the thresholds and the output values according to the techniques given in Section 5.3. In particular the statistical power of f f .k/g, useful in scaling the quantizer characteristic, can be derived from (5.136), assuming known Ms and G p , Mf D Ms Gp (5.138)
Regarding the predictor, once the number of coefcients N is xed, we need to determine the coefcients fci g, i D 1; : : : ; N , that minimize M f . For example in the case N D 1, recalling (2.91), the optimum value of c1 is given by .1/, the correlation coefcient of the input signal at lag 1. Then we have Gp D 1 1 2 .1/ (5.139)
ignoring the effect of the quantizer, that is for fsq .k/ D s.k/g.
389
G p D 1=.1
2 .1// (dB)
Figure 5.56. (a) Reconstruction signal for a DPCM, with (b) a 6 level quantizer.
We give in Table 5.13 the values of G p for three values of .1/. We note that, for an input having .1/ D 0:85, a simple predictor with one coefcient yields a prediction gain equivalent to one bit of the quantizer: consequently, given the total 3q , the DPCM scheme allows us to use a transmission bit rate lower than that of PCM. Evidently, for an input with .1/ D 0 there is no advantage in using the DPCM scheme. O For a simple predictor with N D 1 and c1 D 1, hence s .k/ D sq .k 1/, Figure 5.56a illustrates the behavior of the reconstruction signal after DPCM with the six-level quantizer shown in Figure 5.56b. We note that the minimum level of the quantizer still determines the statistical power of the granular noise in fsq .k/g; the maximum level of the quantizer is instead related to the slope overload distortion in the sense that if Q L=2 is not sufciently large, as shown in Figure 5.56a, the output signal cannot follow the rapid changes of the input signal. In the specic case, being Q L=2 =Tc < max js.k/ s.k 1/j, fsq .k/g presents a slope different from that of fs.k/g in the instants of maximum variation.
5.5.2
Alternative conguration
If we use few quantization levels, the predictor of the scheme of Figure 5.55, having as input fsq .k/g instead of fs.k/g, can give poor performance because of the large quantization
390
Figure 5.57. DPCM scheme with quantizer inserted after the feedback loop: a) encoder, b) decoder.
noise present in fsq .k/g. An alternative consists in using the scheme of Figure 5.57, where the following relations hold. Encoder: f .k/ D s.k/ s .k/ O (5.140) (5.141) (5.142) i/ (5.143)
ci s.k C 1
(5.144) i/ (5.145)
ci sq;o .k C 1
At the encoder, s .k/ is obtained from the input signal without errors. However, the O O prediction signal reconstructed at the decoder is so .k/ 6D s .k/. In fact, from (5.144) and O
391
(5.145), even if by chance so .i/ D s .i/, for i k 1, as f q .k 1/ 6D f .k 1/ then O O sq;o .k 1/ 6D s.k 1/, and consequently so .k/ 6D s .k/. O O A difculty of the scheme is that, depending on the function C.z/, the difference between O the prediction signals, so .k/ s .k/, may be non-negligible. As a result the output O O sq;o .k/ D so .k/ C f .k/ C eq; f .k/ D so .k/ C s.k/ O D s.k/ C [Oo .k/ s s .k/ C eq; f .k/ O s .k/] C eq; f .k/ O (5.146)
can assume values that are quite different from s.k/. Observation 5.4 Note that the same problem mentioned above may occur also in the scheme of Figure 5.55 because of errors introduced by the binary channel, though to a lesser extent as compared to the scheme of Figure 5.57, as the signal f f q .k/g at the encoder is affected by a smaller disturbance. For both congurations, however, the inverse prediction error lter must suppress the propagation of such errors in a short time interval. This is difcult to achieve if the transfer function 1=[1 C.z/] has poles near the unit circle, and consequently the impulse response is very long.
5.5.3
ci sq .k
i/
(5.147)
where sq .k/ is the reconstruction signal, which in the case of a feedback quantizer system is given by sq .k/ D s.k/ C eq; f .k/. For the design of the predictor, we choose the coefcients fci g that minimize the statistical power of the prediction error, M f D E[.s.k/ s .k//2 ] O (5.148)
We introduce the following vectors and matrices. Vector of prediction coefcients c D [c1 ; : : : ; c N ]T Vector of correlation coefcients of fs.k/g D [.1/; : : : ; .N /]T where .i/ is dened in (1.540). (5.150) (5.149)
392
3 7 7 7 7 7 7 7 7 7 5
(5.153)
Recalling the analysis of Section 2.2, the optimum prediction coefcients are given by the matrix equation (2.78) copt D The corresponding minimum value of M f is obtained from (2.79), M f D Ms .1
T copt /
(5.154)
(5.155)
The difculty of this formulation is that to determine the solution we need to know the value of 3q (see (5.153)). We may consider the solution with the quantizer omitted, hence 3q D 1, and depends only on the second order statistic of fs.k/g. In this case some efcient algorithms to determine c and M f in (5.154) and (5.155) are given in Sections 2.2.1 and 2.2.2.
In general it is very difcult to analyze the effects of feq; f .k/g on G p , except in the case N D 1, for which (5.154) becomes 1 D .1/ (5.157) copt;1 1 C 3q Then copt;1 D .1/ 1 C .1=3q / (5.158)
Assuming for fs.k/g and feq; f .k/g the correlations are expressed by (5.27) and (5.28), we get Dividing by Ms we obtain rsq .n/ D rs .n/ C Meq; f n rsq .n/ Ms D .n/ C 3q 1 n (5.151)
(5.152)
393
The above relations show that if 3q is small, that is, if the system is very noisy, then copt;1 is small and G p tends to 1. Only for 3q D 1 it is copt;1 D .1/. It may occasionally happen that a suboptimum value is assigned to c1 : we will try to evaluate the corresponding value of G p . For N D 1 and any c1 it is M f D E[.s.k/ ' Ms .1 c1 .s.k 1/ C eq; f .k 1///2 ] (5.160)
As from (5.135) it follows Meq; f D Ms =3q , where from (5.137) 3q D G p 3q; f , observing (5.160) we obtain Gp D 1 1
2 .c1 =3q; f / 2 2c1 .1/ C c1
(5.161)
hence 3q; f depends only on the number of quantizer levels. Note that (5.161) allows the computation of the optimum value of c1 for a predictor with N D 1 in the presence of the quantizer: however, the expression is complicate and will not be given here. Rather we will derive G p for two values of c1 . 1. For c1 D .1/ we have Gp D where the factor .1 2. For c1 D 1 we have 1 1 1 2 .1/ 2 .1/ 3q; f (5.162)
2 .1/=3q; f / is due only to the presence of the quantizer. 1 1 Gp D 2.1 .1// 1 3q; f (5.163)
We note that the choice c1 D 1 leads to a simple implementation of the predictor: however, this choice results in G p > 1 only if .1/ > 1=2. Various experiments with speech have demonstrated that for very long observations, of the order of one second, the prediction gain for a xed predictor is between 5 and 7 dB, and saturates for N 2; in fact, speech is a non-stationary signal and adaptive predictors should be used.
5.5.4
Adaptive predictors
In adaptive differential PCM (ADPCM), the predictor is time-varying. Therefore we have s .k/ D O
N X i D1
ci .k/sq .k
i/
(5.164)
394
The vector c D [c1 ; : : : ; c N ]T is chosen to minimize M f over short intervals within which the signal fs.k/g is quasi-stationary. Speech signals have slowly-varying spectral characteristics and can be assumed as stationary over intervals of the order of 525 ms. Also for ADPCM two strategies emerge.
395
Figure 5.58. ADPCM scheme with feedforward adaptation of both predictor and quantizer: (a) encoder, (b) decoder.
where 1 1 controls the stability of the decoder if, because of binary channel errors, it occasionally happens fQq .k/ 6D f q .k/. Table 5.14 gives the algorithm, while Figure 5.61 illustrates the implementation; in Q decoding the same equations are used with fQq .k/ in place of f q .k/ and therefore sq .k/ in place of sq .k/.
396
Figure 5.59. (a) Speech level measured on windows of 128 samples, and corresponding prediction gain Gp for: (b) a xed predictor (N D 3), (c) an adaptive predictor (N D 10). For these measurements the quantizer was removed. Table 5.14 Adaptation equations of the LMS adaptive predictor.
Initialization c.0/ D 0 sq .0/ D 0 (or sq .0/ D s.0/) s .0/ D 0 O f .k/ D s.k/ s .k/ O f q .k/ D Q[ f .k/] c.k C 1/ D 1 c.k/C f q .k/sq .k/ O sq .k/ D s .k/ C f q .k/ s .k C 1/ D cT .k C 1/sq .k C 1/ O
For k D 0; 1; : : :
Observation 5.5 For a stationary input, as for example a modem signal, the LMS adaptive prediction can be used to easily determine the predictor coefcients; once the convergence is reached, however, it is better to switch off the adaptation. This observation does not apply to speech, which presents characteristics that may change very rapidly with time.
397
Figure 5.60. ADPCM scheme with feedback adaptation for both predictor and quantizer: (a) encoder, (b) decoder.
398
Performance
Objective and subjective experiments conducted on speech signals sampled at 8 kHz have indicated that adopting ADPCM rather than PCM leads to a saving of 2 to 3 bits in encoding: for example, a 5-bit ADPCM scheme yields the same quality as a 7-bit PCM. Obviously in both cases the quantizer is non-uniform.
5.5.5
In this section we omit the quantizer and we analyze alternative structures for the predictor. For further study on various signal models (AR, MA and ARMA), we refer to Section 1.12.
All-pole predictor
The predictor considered in the previous section implies an inverse prediction error lter or synthesis lter (see Figure 2.9) with transfer function H .z/ D S.z/ D F.z/ 1 1 C.z/ (5.167)
which has only poles (neglecting zeros at the origin). The predictor refers to an input model whose samples are given by s.k/ D We consider two cases.
N X i D1
ai s.k
i/ C w.k/
(5.168)
399
1. fw.k/g is white noise. Then (5.168) implies an AR(N ) model for the input. The PN n and optimum predictor that minimizes E[.s.k/ s .k//2 ] is C.z/ D O nD1 an z it yields f .k/ D w.k/. 2. fw.k/g is a periodic sequence of impulses, w.k/ D A
C1 X nD 1
nP
(5.169)
with P N . In this case (5.168) implies that also fs.k/g is a periodic signal of period PN n and it yields f .k/ D w.k/, P. The optimum predictor is still C.z/ D nD1 an z equal to a periodic sequence of impulses. We observe that the two cases model in a simplied manner the input to an all-pole lter whose output is unvoiced (case 1) or voiced (case 2). The coding scheme that makes use of an all-pole predictor is called linear predictive coding (LPC) and the prediction error f f .k/g is called LPC residual.
All-zero predictor
For an MA input model, the prediction signal is given by s .k/ D O
q X i D1
bi f .k
i/
(5.170)
Correspondingly from (5.124), for sq .k/ D s.k/, the synthesis lter has a FIR all zero transfer function H .z/ D 1 C
q X i D1
bi z
(5.171)
Incidentally we note that an approximate LMS adaptation of the coefcients fbi g is given by bi .k C 1/ D bi .k/ C b f q .k/ f q .k i/ i D 1; : : : ; q (5.172)
Pole-zero predictor
The general case refers to an ARMA( p; q) input model. In this case s .k/ D O Correspondingly, we have 1C H .z/ D 1
q X i D1 p X i D1 p X i D1 q X i D1
ci s.k
i/ C
bi f .k
i/
(5.173)
bi z ci z
(5.174)
i
400
s(k) +
f(k)
f(k)
sq(k)=s(k)
+
^ s(k) b b
+
^ s(k)
+
c sq(k)=s(k)
(a)
(b)
The equations (5.173) and (5.174) are illustrated in Figure 5.62. For the LMS adaptation of the coefcients in (5.173), we refer to (5.166) for the coefcients fci g and to (5.172) for the coefcients fbi g. This conguration was adopted by the ADPCM G.721 standard at 32 kbit/s (see Table 5.16), that uses an LMS adaptive predictor with 2 poles and 4 zeros; the 5 bit quantizer is adapted by the Jayant scheme.
Pitch predictor
An alternative structure exploits the quasi-periodic behavior, of period P, of voiced spurts of speech. In this case it is convenient to use the estimate O s .k/ D s` .k/ C ss .k/ O O where s` .k/ D s.k O P/ (5.176) (5.175)
is the long-term estimate. In (5.176) is the pitch gain, and P is the pitch period expressed in number of samples. Let f f ` .k/g be the corresponding prediction error f ` .k/ D s.k/ Then, in (5.175), ss .k/ D O
N X i D1
s.k
P/
(5.177)
ci f ` .k
i/
(5.178)
is the short-term estimate. The prediction error f .k/ D s.k/ s .k/ D f ` .k/ O
N X i D1
ci f ` .k
i/
(5.179)
401
s(k)
+ -
fl(k)
+ -
f (k) ^ (k) ss
cl
long-term predictor cl (z)= z -P
^ sl (k)
c
short-term predictor c(z)= ci z-i
i=1 N
APC
Because P is usually in the range from 40 to 120 samples, for speech signals sampled at 8 kHz the whole predictor in (5.179) is in fact an all-pole type with a very high order N . The subdivision into two terms, even if not optimum, has the advantage of allowing a very simple computation of the various parameters. 1. Computation of long-term predictor through minimization of the cost function min E[ f ` .k/2 ] D E[.s.k/
;P
s.k
P//2 ]
(5.180)
(5.181)
where .n/ represents the correlation coefcient of fs.k/g at lag n, and D .P/ (5.182)
2. Determination of the short-term predictor through minimization of the cost function: min E[ f 2 .k/]
c
(5.183)
From the estimate of the autocorrelation sequence of fs.k/g, once P and are determined, the autocorrelation sequence of f f ` .k/g is easily computed. Then the coefcients fci g of the long-term predictor can be obtained by solving a system of equations similar to (5.154), where and depend on the autocorrelation coefcients of f f ` .k/g. Experimental measurements, initially conducted by Atal [5], have demonstrated that adapting the various coefcients by the feedforward method every 5 ms we get high quality speech reproduction with an overall bit rate of only 10 kbit/s, using only a one-bit quantizer. The encoder and decoder schemes are given in Figure 5.64: they form the adaptive predictive coding (APC) scheme which differs from the DPCM for the inclusion of the long-term predictor.
402
For voiced speech, the improvement given by the long-term predictor to lowering the LPC residual is shown in Figure 5.65. Without the long-term predictor the LPC residual presents a peak at every pitch period P; as shown in Figure 5.65c, these peaks are removed by the long-term predictor. The frequency-domain representation of the three signals in Figure 5.65 is given in Figure 5.66. We note that plots in Figures 5.66a and 5.66b exhibit some spectral lines, due to the periodic behavior of the corresponding signals in the time domain, whereas these lines are attenuated in the plot of Figure 5.66c.
403
Figure 5.65. (a) Voiced speech, (b) LPC residual, (c) LPC residual with long-term predictor.
Figure 5.66. DFT of signals of Figure 5.65: (a) voiced speech, (b) LPC residual, (c) LPC residual with long-term predictor.
404
Other long-term predictors with 2 or 3 coefcients have been proposed: although they are more effective, the determination of their parameters is much more complicate than the approach (5.180). There are also numerous methods that are more robust and effective than (5.181) to determine the pitch period P. To avoid this very laborious computation, all-pole predictors have been proposed with more than 50 coefcients, thus partly assimilating the long-term predictor in the overall predictor. Two improvements with respect to the basic APC scheme are outlined in the following observations [3]. Observation 5.6 From the standpoint of perception, it is important to have a signal-to-noise ratio that is constant in the frequency domain: this yields the so-called spectral shaping of the error, obtained by ltering the residual error so that it is reduced at frequencies where the signal has low energy and enhanced at frequencies where the signal has high energy. Observation 5.7 In APC, parameters associated with the prediction coefcients fci g i D 1; : : : ; N , are normally sent: for example reection coefcients (PARCOR), area functions or line spectrum pairs (LSP).
5.6
5.6.1
Delta modulation
Oversampling and quantization error
For an input signal s.t/, t 2 <, WSS random process with bandwidth B, let the sampling period be Tc D where T0 D and F0 is the oversampling factor. Let x.k/ D s.kTc / be the sampled version of s.t/, with autocorrelation T0 rx .n/ D rs .nTc / D rs n F0
C1 X `D 1
T0 F0 1 2B
(5.184)
(5.185)
(5.186)
(5.187)
Equation (5.188) is obtained from (5.187) using (1.90) and the denition (1.247).
405
Figure 5.67 shows the effect of oversampling on fx.k/g. In particular, from (5.187) and from Figure 5.67a we note that by increasing F0 the samples fx.k/g become more correlated; moreover, from (5.188) we have that the spectrum of fx.k/g presents images that are more spaced apart from each other. Let us now quantize fx.k/g. With reference to Figure 5.68, let xq .k/ be the quantized signal and eq .k/ the corresponding quantization error xq .k/ D x.k/ C eq .k/ (5.189)
Figure 5.67. Effects of oversampling for two values of the oversampling factor F0 .
406
For feq .k/g white with statistical power Meq , depending only on the number of quantizer levels (see (5.44) and (5.98)), we have Peq . f / D Meq T0 F0 (5.190)
Consequently, by increasing F0 the PSD of eq decreases in amplitude. Moreover, from (5.189), for the non-correlation assumption between x and eq , we have Pxq . f / D Px . f / C Peq . f / (5.191)
From (5.189), by ltering fxq .k/g with an ideal lowpass lter g having bandwidth B and unit gain, at the output we will have yq .k/ D x.k/ C eq;o .k/ where eq;o .k/ D eq g.k/ has PSD given by Peq;o . f / D Meq Then Meq;o D Meq F0 (5.194) T0 f rect F0 2B (5.193) (5.192)
and, with reference to (5.192), the effective signal-to-noise ratio is given by 3q;o D 3q F0 (5.195)
where 3q D Mx =Meq is to the signal-to-noise ratio after the quantizer. In conclusion, under the assumption that the quantization noise is white, oversampling improves performance by a factor F0 . However, the encoder bit rate, Rb D b 1 F0 Db Tc T0 (5.196)
increases proportionally to F0 : for example, F0 D 4 improves 3q by 6 dB, but at the expense of quadrupling the bit rate. Therefore oversampling by large factors is used only locally before PCM or in compact disk (CD) applications to simplify the analog interpolation lter at the receiver.
407
5.6.2
Linear delta modulation is a DPCM scheme with oversampled input signal, 1 2B Tc (5.197)
and a quantizer with only two levels (b D 1). Then the encoder bit rate is equal to the sampling rate, Rb D 1 (bit/s) Tc (5.198)
The high value of F0 implies a high predictability of the input sequence: therefore a predictor with a few coefcients gives a high prediction gain and the quantizer can be reduced to the simplest case of b D 1. We note, moreover, that one-bit code words eliminate the need for framing of the code words at the transmitter and at the receiver, thus simplifying the overall system. For a predictor with only one coefcient c1 , the coding scheme, which is illustrated in Figure 5.69, is called a linear delta modulator (LDM). The following relations hold.
408
Encoder: f .k/ D s.k/ ( f q .k/ D s .k/ O 1 .c.k/ D 1/ if f .k/ 0 1 .c.k/ D 0/ if f .k/ < 0 (5.199) (5.200) (5.201)
sq .k/ D c1 sq .k
The system is based on three parameters (1=Tc , 1, and c1 ), that are appropriately selected. For example the choice of c1 D 1 considerably simplies (5.201) and (5.203), that become simple accumulator expressions. However, typically we set c1 1 so that random transmission errors do not propagate indenitely in the reconstruction signal (5.203).
LDM implementation
Digital implementation. An implementation of the scheme of Figure 5.69 for c1 D 1 is given in Figure 5.70a. Letting b.k/ D sgn[ f .k/], this implementation involves the accumulation of fb.i/g, i k, by an up-down counter (ACC). The accumulated value is proportional to sq .k/. Mixed analog-digital implementation. An alternative to the previous scheme, which requires carrying out the operation f .k/ D s.k/ s .k/ in the digital domain, is obtained O by placing a DAC after the accumulator and carrying out the comparison in the analog domain, as illustrated in Figure 5.70b. Note that the decoder consists simply of an accumulator followed by a DAC, which performs the function of the lter g, to eliminate the out-of-band noise, and of the gain 1 of the quantizer step size. Analog implementation. In many applications it is convenient to implement the analog accumulator by an integrator: thus we obtain the implementation of Figure 5.70c, where the DAC is often a simple holder. At the receiver, the integrator has a bandwidth B equal to that of the input signal.
409
granular noise
s(k) s(k-1)
slope overload distortion
sq (k) ^ s(k)
} fq(k)
t
-1 +1 +1
{
0 Tc +1 +1 +1 +1 +1 -1 +1 -1
b(k)
410
distortion a large 1 is needed, so that d 1 s.t/ (5.205) max t dt Tc and the reconstruction signal fsq .k/g can follow very rapid changes of the input signal. In other words, for a given value of maxt j.d=dt/s.t/j, if we reduce 1 to decrease the granular noise we must also reduce Tc to limit the slope overload distortion; as a consequence we have that LDM requires a very high oversampling factor F0 to give satisfactory performance. We note that doubling the sampling rate we obtain an increment in 3q;o of 9 dB: 3 dB are due to ltering of the out-of-band noise and 6 dB to the reduction of the granular noise, as 1 can be halved. In speech applications with a bandwidth of about 3 kHz, to have a 3q;o of approximately 35 dB we need F0 33, which requires a sampling rate of the order of 200 kHz. The optimum value of 1 is given approximately by p (5.206) 1opt D 2Ms .1 x .1// ln.2F0 / where x .1/ D rs .Tc /=rs .0/.
5.6.3
To reduce both granular noise and slope overload distortion, the only possibility in the LDM is to reduce Tc . An alternative is represented by an adaptive scheme for the step size 1, as shown in Figure 5.72. In particular, the Jayant algorithm uses the following relation 1.k/ D p1.k 1/ (5.207)
411
where pD
po > 1 pg < 1
(5.208)
We note that in this scheme 1.k/ also depends on c.k/. The following relations between the signals of Figure 5.72 hold. Encoder: f .k/ D s.k/ s .k/ O (5.209) b.k/ D sgn f .k/ 1.k/ D p1.k 1/ f q .k/ D 1.k/b.k/ s .k C 1/ D c1 [O .k/ C f q .k/] O s Decoder: 1.k/ D p1.k 1/ (5.214) f q .k/ D 1.k/b.k/ (5.215) (5.216) sq .k/ D c1 sq .k 1/ C f q .k/ sq;o .k/ D sq g.k/ (5.217) Typical values for po and pg are given by 1:25 < po < 2 and po pg 1. A graphic representation of ADM encoding is shown in Figure 5.73. Experiments on speech signals show that by doubling the sampling rate in the ADM we get an improvement of 10 dB in 3q . In some applications ADM encoding is preferred to PCM because of its simple implementation, in spite of the higher bit rate. (5.210) (5.211) (5.212) (5.213)
(5.218)
granular noise
^ s(k)
Tc
b(k)
+1 +1 +1 -1 +1 -1 +1 -1 +1 -1
412
where 0 < 1, and D1 and D2 are suitable positive parameters with D2 D1 . The value controls the speed of adaptation. The main difculty of this scheme is the sensitivity of fsq .k/g to transmission errors, especially for D 1; choosing < 1 mitigates the effects of transmission errors, at the expense of worse performance.
5.6.4
We consider an alternative scheme to the three implementations of Section 5.2.2 to generate a linear PCM encoded signal, that employs the LDM implementation of Figure 5.70c, as illustrated in Figure 5.74. It is sufcient to accumulate fb.k/g to obtain a PCM representation of the input s.t/; using a decimator lter, that is a lowpass lter followed by a downsampler, we obtain the PCM output signal sampled at the minimum rate 1=T0 . Disregarding the ltering effect on noise, that brings a gain of 10 log10 F0 dB, we observe that to generate a PCM signal fc PC M .k/g with an accuracy of b bits, the oversampling factor F0 must be at most equal to 2b and, in general, 1 F0 2b For example, for b D 8 and 1=T0 D 8 kHz, this means 1=Tc D F0 =T0 ' 2 MHz. (5.222)
413
5.6.5
With reference to the scheme of Figure 5.70c, to enhance the low frequency components of speech signals we can insert a pre-emphasis integrator before the LDM encoder. A differentiator then has to be inserted at the LDM decoder, which simplies the LDM integrator: therefore the decoder becomes a simple lowpass lter. Thus we get the general scheme of Figure 5.75a and the simplied scheme of Figure 5.75b: we note that the DACs are simple holders of binary signals. It is interesting to observe the simplicity of the 6DM implementation. Moreover, recalling that the spectrum of quantization noise in PCM and LDM is at, 6DM presents the advantage that the noise is colored and for the most part is found outside the passband of the desired signal. Therefore it can be removed to a large extent by a simple lowpass lter. One of the most frequent applications of 6DM is in linear PCM encoders where, similarly to the scheme of Figure 5.74, it is sufcient to employ a 6DM followed by a digital decimator lter with input the binary signal fb.k/g. Note that, with respect to the scheme of Figure 5.74, the accumulator has been removed.
5.7
Coding by modeling
In the coding schemes investigated so far, PCM, DPCM, and their variations, the objective is to reproduce at the decoder a waveform that is as close as possible to the input signal. We now take a different approach and, given the general coding scheme of Figure 5.76,
414
the source fs.k/g, for example speech, is modeled by an AR.N / linear system H .z/ D 1
N X i D1
(5.223) ci z
i
with input f f .k/g. In (5.223) the coefcients fci g and are obtained by the prediction algorithms of Section 2.2. In particular, as we will assume that f f .k/g has unit statistical power, the standard deviation in (5.223) is given by p (5.224) D JN where J N is the statistical power of the prediction error. The difference among the various coding schemes consists in the form of excitation. Three examples follow. Regular pulse excited (RPE). The excitation signal consists of a train of undersampled impulses, derived from the residual signal. Multipulse LP (MELP). The excitation signal consists of a certain number of impulses with suitable amplitude and lag. Codebook excited linear prediction (CELP). The excitation signal is selected by a collection of possible waveforms stored in a table. We will now analyze in detail some coding schemes. For further study we refer the reader to [3, 6].
Vocoder or LPC
The general scheme for the conventional LPC, known also as the LPC vocoder, is illustrated in Figure 5.77. At the encoder, the signal is classied as voiced or unvoiced, and the LCP parameters are extracted, together with the pitch period P for the voiced case; however,
415
the prediction residual error is not transmitted. At the decoder, for the voiced case a train of impulses with period P is produced, whereas for the unvoiced case white noise is produced. The excitation is then ltered by the AR lter to generate the reconstruction signal. In an early LPC scheme for military radio applications (LPC-10), the input signal sampled at 8 kHz is segmented into blocks of 180 samples. For the analysis of the LPC parameters the covariance method is used; overall 54 bits per block are needed with a bit rate of 2400 bit/s.
RPE coding
The RPE coding scheme, illustrated in Figure 5.78a, is a particular case of residual excited LP (RELP) coding in which the excitation is obtained by downsampling the prediction residual error by a factor of 3, as shown in Figure 5.78b; the excitation sequence is then quantized using a 3-bit adaptive non-uniform quantizer. The choice of the best of the three subsequences (actually four are used in practice) is made by the analysis-by-synthesis (ABS) approach, where all the excitations are tried: the best is that which produces the output closest to the original signal. The standard ETSI for GSM (06.10) includes also a long-term predictor as the one of Figure 5.64. The bit rate is 13 kbit/s, with a latency lower than 80 ms, operating with blocks of 160 samples.
416
CELP coding
As shown in Figure 5.79, the excitations belong to a codebook obtained in a random way, or by vector quantization (see Section 5.8) of the residual signal. The choice of the excitation (index of the codebook) is made by the ABS approach, trying to minimize the output of the weighting lter; also in this case the predictor includes a long term component.
417
Multipulse coding
It is similar to CELP coding, with the difference that the minimization procedure is used to determine the position and amplitude of a specic number of impulses. The analysis procedure is less complex than that of the CELP scheme.
5.8
Vector quantization (VQ) is introduced as a natural extension of the scalar quantization (SQ) concept. However, using multidimensional signals opens the way to many techniques and applications that are not found in the scalar case [7, 8]. The basic concept is that of associating with an input vector s D [s1 ; : : : ; s N ]T , generic sample of a vector random process s.k/, a reproduction vector sq D Q[s] chosen from a nite set of L elements (code vectors), A D fQ1 ; : : : ; Q L g, called codebook, so that a given distortion measure d.s; Q[s]/ is minimized. Figure 5.80 exemplies the encoder and decoder functions of a VQ scheme. The encoder computes the distortion associated with the representation of the input vector s by each
418
reproduction vector of A and decides for the vector Qi of the codebook A that minimizes it; the decoder associates the vector Qi to the index i received. We note that the information transmitted over the digital channel identies the code vector Qi : therefore it depends only on the codebook size L and not on N , dimension of the code vectors. An example of input vector s is obtained by considering N samples at a time of a N C 1/Tc /]T , or the N LPC coefcients, speech signal, s.k/ D [s.k N Tc /; : : : ; s..k N T , associated with an observation window of a signal. s.k/ D [c1 .k/; : : : ; c N .k/]
5.8.1
Characterization of VQ
Considering the general case of complex-valued signals, a vector quantizer is characterized by Source or input vector s D [s1 ; s2 ; : : : ; s N ]T 2 C N . Codebook A D fQi g, i D 1; : : : ; L, where Qi 2 C Distortion measure d.s; Qi /. Quantization rule (minimum distortion) Q:C
N N
is a code vector.
! A with
Qi D Q[s]
(5.225)
Denition 5.1 (Partition of the source space) The equivalence relation Q D f.s1 ; s2 / : Q[s1 ] D Q[s2 ]g (5.226) which associates input vector pairs having the same reproduction vector, identies a partition R D fR1 ; : : : ; R L g of the source space C N , whose elements are the sets R` D fs 2 C
N
: Q[s] D Q` g
` D 1; : : : ; L
(5.227)
The sets fR` g, ` D 1; : : : ; L, are called Voronoi regions. It can be easily demonstrated that the sets fR` g are non-overlapping and cover the entire space C N :
L [ `D1
R` D C
Ri \ R j D ;
8i 6D j
(5.228)
In other words, as indicated by (5.227) every subset R` contains all input vectors associated by the quantization rule with the code vector Q` . An example of partition for N D 2 and L D 4 is illustrated in Figure 5.81.
419
R4
R1
1 0
Q4
C2
Q 11 00
1
Q 1 0
Q 11 00
R2
R3
Figure 5.81. Partition of the source space C 2 in four subsets or Voronoi regions.
Rate per dimension RI D Rate in bit/s Rb D RI log2 L D (bit/s) Tc N Tc (5.231) Rq log2 L D (bit/sample) N N (5.230)
where Tc denotes the time interval between two consecutive samples of a vector. In other words, in (5.231) N Tc is the sampling period of the vector sequence fs.k/g. Distortion d.s; Qi / The distortion is a non-negative scalar function of a vector variable, d:C
N
(5.232)
A ! <C
(5.233)
If the input process s.k/ is stationary and the probability density function ps .a/ is known, we can compute the mean distortion as D.R; A/ D E[d.s; Q[s]/] D
L X `D1 L X `D1
(5.234) (5.235)
(5.236)
420
where N d` D
Z
R`
(5.237)
K !1
(5.238)
In practice we always assume that the process fsg is stationary and ergodic, and we use the average distortion (5.238) as an estimate of the expectation (5.234). Dening Qi D [Q i;1 ; Q i;2 ; : : : ; Q i;N ]T we give below two measures of distortion of particular interest. 1. Distortion as the ` norm to the -th power: " N X d.s; Qi / D jjs Qi jj D jsn
nD1
(5.239)
Q i;n j
#= (5.240)
The most common version is the squared distortion:11 d.s; Qi / D jjs 2. ItakuraSaito distortion: d.s; Qi / D .s Qi / H Rs .s Qi / D
N N XX nD1 mD1
Qi jj2 D 2
N X nD1
jsn
Q i;n j2
(5.241)
.sn
Q i;m / (5.242)
where Rs is the autocorrelation matrix of the vector s .k/, dened in (1.346), with elements [Rs ]n;m , n; m D 1; 2; : : : ; N .
11 Although the same symbol is used, the metric dened by (5.241) is the square of the Euclidean distance (1.38).
421
where F.N / is the space lling gain. In the scalar case the partition regions must necessarily be intervals. In an N dimensional space, Ri can be shaped very closely to a sphere. The asymptotic value for N ! 1 equals F.1/ D 2 e=12 D 1:4233 D 1:53 dB. S.N / is the gain related to the shape of ps .a/, dened as12 S.N / D jj ps .a/jj1=3 Q jj ps .a/jj N =.N C2/ Q (5.245)
where ps .a/ is the probability density function of the input s considered with uncorQ related components. S.N / does not depend on the variance of the random variables of s, but only on the norm order N =.N C 2/ and shape ps .a/. For N ! 1, we Q obtain S.1/ D M.N / is the memory gain, dened as M.N / D jj ps .a/jj N =.N C2/ Q jj ps .a/jj N =.N C2/ (5.247) jj ps .a/jj1=3 Q jj ps .a/jj1 Q (5.246)
where ps .a/ is the probability density function of the input s. The expression of M.N / depends on the two functions ps .a/ and ps .a/, which differ for the correlation Q among the various vector components; obviously if the components of s are statistically independent, we have M.N / D 1; otherwise M.N / increases as the correlation increases.
5.8.2
Optimum quantization
Our objective is to design a vector quantizer, choosing the code vectors of the codebook A and the partitioning R so that the mean distortion given by (5.234) is minimized. Two necessary conditions arise. Rule A (Optimum partition). Assuming the codebook A D fQ1 ; : : : ; Q L g xed, we want to nd the optimum partition R that minimizes D.R; A/. Observing (5.234) the solution is given by Ri D fs : d.s; Qi / D min d.s; Q` /g
Q` 2A
i D 1; : : : ; L
(5.248)
jj ps .a/jj= D Q
ps .a1 ; : : : ; a N / da1 : : : da N Q
= (5.244)
422
Rule B (Optimum codebook). Assuming the partition is R given, we want to nd the optimum codebook A. By minimizing (5.236) we obtain the solution Qi : E[d.s; Qi / j s 2 Ri ] D min E[d.s; Q j / j s 2 Ri ]
Q j 2 Ri
(5.249)
In other words Qi coincides with the centroid of the region Ri . As a particular case, choosing the squared distortion (5.241), (5.237) becomes Z jjs Qi jj2 ps .a/ da 2 1 Ri N Z di D (5.250) ps .a/ da
Ri
423
2. Using the rule A we determine the optimum partition R[A j ] using the codebook A j . 3. We evaluate the distortion associated with the choice of A j and R[A j ] using (5.236). 4. If Dj
1
Dj
Dj
<
(5.252)
we stop the procedure, otherwise we update the value of j. 5. Using rule B we evaluate the optimum codebook associated to the partition R[A j 6. We go back to step 2.
1 ].
424
The solution found is at least locally optimum; nevertheless, given that the number of locally optimum codes can be rather large, and some of the locally optimum codes may give rather poor performance, it is often advantageous to provide a good codebook to the algorithm to start with, as well as trying different initial codebooks. The algorithm is clearly a generalization of the Lloyd algorithm given in Section 5.3.2: the only difference is that the vector version begins with a codebook (alphabet) rather than with an initial partition of the input space. However, the implementation of this algorithm is difcult for the following reasons. The algorithm assumes that ps .a/ is known. In the scalar quantization it is possible, in many applications, to develop an appropriate model of ps .a/, but this becomes a more difcult problem with the increase of the number of dimensions N : in fact, the identication of the distribution type, for example, Gaussian or Laplacian, is no longer sufcient, as we also need to characterize the statistical dependence among the elements of the source vector. The computation of the input space partition is much harder for the VQ. In fact, whereas in the scalar quantization the partition of the real axis is completely specied by a set of .L 1/ points, in the two-dimensional case the partition is specied by a set of straight lines, and for the multi-dimensional case to nd the optimum solution becomes very hard. For VQ with a large number of dimensions, the partition becomes also harder to describe geometrically. Also in the particular case (5.251), the calculation of the centroid is difcult for the VQ, because it requires evaluating a multiple integral on the region Ri .
5.8.3
LBG algorithm
An alternative approach led Linde, Buzo, and Gray [10] to consider some very long realizations of the input signal and to substitute (5.234) with (5.238) for K sufciently large. The sequence used to design the VQ is called training sequence (TS) and is composed of K vectors fs.m/g The average distortion is now given by DD
K 1 X d.s.k/; Q[s.k/]/ K kD1
m D 1; : : : ; K
(5.253)
(5.254)
and the two rules to minimize D become: Rule A Ri D fs.k/ : d.s.k/; Qi / D min d.s.k/; Q` /g
Q` 2A
i D 1; : : : ; L
(5.255)
425
1 X d.s.k/; Q j / m i s.k/2R
i
(5.256)
where m i is the number of elements of the TS that are inside Ri . Using the structure of the Lloyd algorithm with the new cost function (5.254) and the two new rules (5.255) and (5.256), we arrive at the LBG algorithm. Before discussing the details, it is worthwhile pointing out some aspects of this new algorithm. It converges to a minimum, which is not guaranteed to be a global minimum, and generally depends on the choice of the TS. It does not require any stationarity assumption. The partition is determined without requiring the computation of expectations over C N. The computation of Qi in (5.256) is still burdensome. However, for the squared distortion (5.241) we have d.s.k/; Qi / D and (5.256) simply becomes Rule B Qi D 1 X s.k/ m i s.k/2R
i
N X nD1
jsn .k/
Q i;n j2
(5.257)
(5.258)
that is Qi coincides with the arithmetic mean of the TS vectors that are inside Ri .
426
Let A j D fQ1 ; : : : ; Q L g be the codebook at iteration j-th. The splitting procedure generates 2L N -dimensional vectors yielding the new codebook A jC1 D fA j g [ fAC g j where A j D fQi AC D fQi j Typically is the zero vector, D0 and 1 C D 10 so that jj C jj2 0:01 Ms 2 where Ms is the power of the TS. (5.264) r Ms 1 N (5.263) (5.262) g C g i D 1; : : : ; L ; i D 1; : : : ; L (5.260) (5.261) (5.259)
427
the largest contribution to the distortion; then we compute the new partition and proceed in the usual way. We give some practical rules, taken from [3] for LPC applications,13 that can be useful in the design of a vector quantizer. If K =L 30, there is a possibility of empty regions, where we recall K is the number of vectors of the TS, and L is the number of code vectors. If K =L 600, an appreciable difference between the distortion calculated with the TS and that calculated with a new sequence may exist. In the latter situation, it may in fact happen that, for a very short TS, the distortion computed for vectors of the TS is very small; the extreme case is obtained by setting K D L, hence D D 0. In this situation, for a sequence different from the TS (outside TS) the distortion is
13 These rules were derived in the VQ of LPC vectors. They can be considered valid in the case of strongly
correlated vectors.
428
Figure 5.86. Values of the distortion as a function of the number of vectors K in the inside and outside training sequences.
in general very high. As illustrated in Figure 5.86, only if K is large enough, does the TS adequately represent the input process and no substantial difference appears between the distortion measured with vectors inside or outside TS [10].14 Finally we nd that the LBG algorithm, even though very simple, requires numerous computations. We consider, for example, as vector source the LPC coefcients with N D 10, computed over windows of duration equal to 20 ms of a speech signal sampled at 8 kHz. Taking L D 256 we have a rate Rb D 8 bit/20 ms equal to 400 bit/s. As a matter of fact, the
14 This situation is similar to that obtained by the LS method (see Section 3.2).
429
LBG algorithm requires a minimum K D 600 256 roughly corresponds to three minutes of speech.
5.8.4
Variants of VQ
Tree search VQ
A random VQ, determined according to the LBG algorithm, requires: a large memory to store the codebook; a large computational complexity to evaluate the L distances for encoding. A variant of VQ that requires a lower computational complexity, at the expense of a larger memory, is the tree search VQ. As illustrated in Figure 5.87, whereas in the memoryless VQ case the comparison of the input vector s must occur with all the elements of the codebook, thus determining a full search, in the tree search VQ we proceed by levels: rst, we compare s with Q A1 and Q A2 , then we proceed along the branch whose node has a representative vector closest to s. To determine the code vectors at different nodes, for a binary tree the procedure consists of the following steps. 1. Calculate the optimum quantizer for the rst level by the LBG algorithm; the codebook contains 2 code vectors. 2. Divide the training sequence into subsequences relative to every node of level n (n D 2; 3; : : : ; N L E V , N L E V D log2 L); in other words, collect all vectors that are associated with the same code vector. 3. Apply the LGB algorithm to every subsequence to calculate the codebook of level n.
430
Computation of d.; / full search tree search for Rq D 10 (bit/vector) full search tree search 2 Rq 2Rq Computation of d.; / 1024 20
i i D1 2
As an example, the memory requirements and the number of computations of d.; / are shown in Table 5.15 for a given value of Rq (bit/vector) in the cases of full search and tree search. Although the performance is slightly lower, the computational complexity of the encoding scheme for a tree search is considerably reduced.
Multistage VQ
The multistage VQ technique presents the advantage of reducing both the encoder computational complexity and the memory required. The idea consists in dividing the encoding procedure into successive stages, where the rst stage performs quantization with a codebook with a reduced number of elements. Successively, the second stage performs quantization of the error vector e D s Q[s]: the quantized error gives a more accurate representation of the input vector. A third stage could be used to quantize the error of the second stage and so on. We compare the complexity of a one-stage scheme with that of a two-stage scheme, illustrated in Figure 5.88. Let Rq D log2 L be the rate in bit/vector for both systems and assume that all the code vectors have the same dimension N D N1 D N2 . Two-stage: Rq D log2 L 1 C log2 L 2 , hence L 1 L 2 D L. Computations of d.; / for encoding: L 1 C L 2 . Memory: L 1 C L 2 locations. One-stage: Rq D log2 L. Computations of d.; / for encoding: L. Memory: L locations. The advantage of a multistage approach in terms of cost of implementation is evident, however, it has lower performance than a one-stage VQ.
Product code VQ
The input vector is split into subvectors that are quantized independently, as illustrated in Figure 5.89.
431
432
This technique is useful if a) there are input vector components that can be encoded separately because of their different effects, e.g., prediction gain and LPC coefcients, or b) the input vector has too large a dimension to be encoded directly. It presents the disadvantage that it does not consider the correlation that may exist between the various subvectors, that could bring about a greater coding efciency. A more general approach is the sequential search product code, [7], where the quantization of the subvector n depends also on the quantization of previous subvectors. With reference to Figure 5.89, assuming L D L 1 L 2 and N D N1 C N2 , we note that the rate per dimension for the VQ is given by log2 L 1 log2 L 2 log2 L D C N N N whereas for the product code VQ it is given by Rq D Rq D log2 L 1 log2 N2 C N1 N2 (5.265)
(5.266)
5.9
We briey discuss two other coding techniques along with the perceptive aspects related to the hearing apparatus. For further details we refer the reader to [3, 6].
433
5.10
Source coding
We briey mention an important topic, namely source coding, which is used to compress digital information messages. In fact, a discrete-time, discrete-valued source signal can be encoded with a lower average bit rate by means of entropy coding [4], which assigns code words of variable lengths to possible input patterns, i.e. to highly probable input patterns are assigned shorter code words and vice versa. We cite the LempelZiv algorithm as one of the most common source coding algorithms.
434
5.11
We conclude this chapter by giving in Table 5.16 a partial list of the various standards to code audio and speech signals [11]. The rst nine are for narrowband speech applications (see Table 5.1). It is interesting to observe the various standards that adopt CELP coding, listed in Table 5.17: we notice that most of them are for cellular radio applications. It is also interesting to compare the various standards to code video signals given in Table 5.18 with those of Table 5.16 for speech and audio.
Description PCM at 64 kbit/s ADPCM at 32 kbit/s ADPCM at 24 and 40 kbit/s G.723+G.721 embedded ADPCM at 40, 32, 24 and 16 kbit/s (embedded means that a code also includes those of lower rate) LD-CELP at 16 kbit/s (LD stands for low delay) CS-ACELP at 8 kbit/s CS-ACELP at 8 kbit/s with reduced complexity MPC-MLQ at 5.3 and 6.4 kbit/s SBC+ADPCM for wide band speech at 64, 56 and 48 kbit/s . A SBC scheme having two bands, 0 4 kHz and 4 8 kHz is used; in each band there is a G.721 encoder. Bit allocation in the two bands is dynamic, for example, 5+3 or 6+2 VSELP at 7.95 kbit/s (VSELP stands for vector sum excited linear prediction) LPC at 2.4 kbit/s CELP at 4.8 kbit/s RPE-LTP at 13 kbit/s (LTP stands for long-term prediction); there is also a 5.6 kbit/s version SBC at 192 kbit/s per audio channel (stereo) [generally 32 448 kbit/s total] SBC at 128 kbit/s per audio channel [generally 32 384 kbit/s total] SBC+MDCT+Huffman coding at 96 kbit/s per audio channel [generally 32 320 kbit/s total] SBC+MDCT coding at 64 kbit/s per audio channel
6 7 8 9 10
11 12 13 14 15 16 17 18
IS-54 (TIA) FS-1015 (LPC-10E) FS-1016 GSM-FR MPEG1, Layer I MPEG1, Layer II MPEG1, Layer III MPEG2, AAC
5. Bibliography
435
Body ITU ITU ITU TIA TIA TIA/ETSI ETSI U.S. (DoD) U.S. (DoD)
Abbreviations G.723 G.728 G.729 IS-54 IS-95 US-1 GSM-HR FS 1016 MELP
Bit rate 5.27 and 6.3 kbit/s; coding of audio in multimedia systems 16 kbit/s 8 kbit/s; encoding of speech and data 7.95 kbit/s; full rate for North America cellular systems based on D-AMPS 1.2, 2.4, 4.8 and 9.6 kbit/s; coding for North America cellular systems based on CDMA 12.2 kbit/s; enhanced full rate for GSM 5.6 kbit/s; half rate for GSM 4.8 kbit/s 2.4 kbit/s
Application
ISDN Video Telephone 64128 kbit/s ISDN Video Conferencing 128 kbit/s MPEG1 CD-Rom Video 1.5 Mbit/s MPEG2 TV (Broadcast Quality) 6 Mbit/s HDTV (Broadcast Quality) 24 Mbit/s TV (Studio Quality, Compressed) 34 Mbit/s HDTV (Studio Quality, Compressed) 140 Mbit/s TV (Studio Quality, Uncompressed) 216 Mbit/s HDTV (Studio Quality, Uncompressed) 1 Gbit/s
Bibliography
[1] L. R. Rabiner and R. W. Schafer, Digital processing of speech signals. Englewood Cliffs, NJ: Prentice-Hall, 1978. [2] IEEE Signal Processing Magazine, Sept. 1997. vol. 14. [3] D. Sereno and P. Valocchi, Codica numerica del segnale audio. LAquila: Scuola Superiore G. Reiss Romoli, 1996. [4] N. S. Jayant and P. Noll, Digital coding of waveforms. Englewood Cliffs, NJ: PrenticeHall, 1984.
436
[5] B. S. Atal and J. R. Remde, A new model of LPC excitation for producing naturalsounding speech at low bit rates, in Proc. ICASSP, pp. 614617, 1982. [6] B. S. Atal, V. Cuperman, and A. Gersho, eds, Advances in speech coding. Boston, MA: Kluwer Academic Publishers, 1991. [7] A. Gersho and R. M. Gray, Vector quantization and signal compression. Boston, MA: Kluwer Academic Publishers, 1992. [8] R. M. Gray, Vector quantization, IEEE ASSP Magazine, vol. 1, pp. 429, Apr. 1984. [9] T. D. Lookabaugh and R. M. Gray, Highresolution quantization theory and the vector quantizer advantage, IEEE Trans. on Information Theory, vol. 35, pp. 1020 1033, Sept. 1989. [10] Y. Linde, A. Buzo, and R. M. Gray, An algorithm for vector quantizer design, IEEE Trans. on Communications, vol. 28, pp. 8495, Jan. 1980. [11] IEEE Communication Magazine, Sept. 1997. vol. 35.
Chapter 6
Modulation theory
The term modulation indicates the process of translating the information generated by a source into a signal that is suitable for transmission over a physical channel. In the case of digital transmission, the information is represented by a sequence of binary data (bits) generated by the source, or by a digital encoder of analog signals (see Chapter 5). The mapping of a digital sequence to a signal is called digital modulation, and the device that performs the mapping is called digital modulator. A modulator may employ a set of M D 2 waveforms to generate a signal (binary modulation), or, in general, M 2 waveforms (M-ary modulation). The transmission medium determines the channel characteristics, as discussed in Chapter 4; it may consist for example of a twisted-pair cable, a coaxial cable, an optical ber, an infrared link, a radio link, or a combination of them; in any case the channel modies the transmitted waveform by introducing for example distortion, interference, and noise. In this chapter we assume that the channel introduces only additive white Gaussian noise (AWGN). We postpone the study of other effects to the next chapters; only in Section 6.12 we will give some simple results for a channel affected by at fading. The task of the receiver is to detect which signal was transmitted, based on the received signal. Using the vector representation of signals discussed in Section 1.2, in this chapter we will introduce the optimum receiver, referring to the detection theory, and present a survey of the main modulation techniques, e.g., PAM, PSK, QAM, orthogonal, and biorthogonal. The performance of each modulation-demodulation method is evaluated with reference to the bit error probability, and a comparison of the various methods is given in terms of spectral efciency and required transmission bandwidth. The transmission rates achievable by the various modulation methods over a specic channel for a given target error probability are then compared with the Shannon bound, which indicates the maximum rate that can be achieved for reliable transmission.
6.1
We consider rst the transmission of an isolated pulse. With reference to the system illustrated in Figure 6.1, the transmitter generates randomly one of M real-valued waveforms sn .t/, n D 1; : : : ; M, and sends it over the channel; the waveform is corrupted by real-valued additive white Gaussian noise w having zero mean and spectral density N0 =2. The variable a0 in Figure 6.1 is modeled as a discrete r.v.
438
with values in f1; : : : ; Mg, and represents the index, or symbol, of the transmitted signal. Assuming that the waveform with index m is transmitted, that is a0 D m, the received, or observed, signal is given by r.t/ D sm .t/ C w.t/ The receiver, based on r.t/, must decide which among the M hypotheses Hn : r.t/ D sn .t/ C w.t/ n D 1; 2; : : : ; M (6.2) (6.1)
is the most probable, and correspondingly must select the detected value a0 [1]. The theory O exposed in this section can be immediately extended to the case of complex-valued signals. We represent time-continuous signals using the vector notation introduced in Section 1.2. Let f i .t/g, i D 1; : : : ; I , be a complete basis for the M signals fsm .t/g, m D 1; : : : ; M; let sm be the vector representation of sm .t/, t 2 <, sm .t/ where smi D hsm ; Z
ii C1 1
! sm D [sm1 ; : : : ; sm I ]T
(6.3)
sm .t/
i .t/ dt
m D 1; : : : ; M
i D 1; : : : ; I
(6.4)
Recall that the set fsm g, m D 1; : : : ; M, is also called the system constellation. The basis of I functions may be incomplete for the representation of the noise signal w. In any case, we express the noise as w.t/ D w .t/ C w? .t/ where w .t/ D and where wi D hw; Z
ii D 1 C1 I X i D1
(6.5)
wi
i .t/
(6.6)
w.t/
i .t/ dt
(6.7)
439
In other words, w is the component of w that lies in the space spanned by the basis f i .t/g, t 2 <, i D 1; : : : ; I , whereas w? is the error due to this representation. Since w? .t/ D w.t/ w .t/ (6.8)
is orthogonal to f i .t/g, i D 1; : : : ; I , and hence also to w , we can say that w? lies outside of the desired signal space and, as we will state later using the theorem of irrelevance, it can be ignored because it is irrelevant for the detection. The vector representation of the component of the noise signal that lies in the span of f i .t/g is given by w D [w1 ; : : : ; w I ]T D w (6.9)
E[w.t/]
i .t/ dt
D0
i D 1; : : : ; I
(6.10)
N0 .t1 2
t2 /
(6.11)
N0 2
(6.12)
i .t/ j .t/ dt
N0 D 2 D
N0 i 2
i; j D 1; : : : ; I
because of the orthogonality of the basis f i .t/g. Hence the components fwi g are uncorrelated. As fwi g, i D 1; : : : ; I , are jointly Gaussian uncorrelated random variables with zero mean, then they are statistically independent with equal variance given by I2 D N0 2 (6.13)
440
Sufcient statistics
Dening r D [r1 ; : : : ; r I ]T with ri D hr;
ii
(6.14)
the components of the vector r are called sufcient statistics 1 to decide among the M hypotheses. Therefore we get the formulation equivalent to (6.2), Hn : r D sn C w n D 1; : : : ; M (6.15) From the above results, the probability density function of r, under the hypothesis that waveform n is transmitted, is given by:2 1 1 2 <I prja0 . j n/ D r exp jj sn jj2 (6.16) !I N0 N0 2 2
Decision criterion
We subdivide the space < I of the received signal r into M non-overlapping regions Rn SM ( nD1 Rn D < I and Rn \ Rm D ; for n 6D m). Then we adopt the following decision rule: O choose Hn (and a0 D n) if r 2 Rn (6.17) The choice of M regions is made so that the probability of a correct decision is maximum. Let pn D P[a0 D n] be the transmission probability of the waveform n, or a priori probability. Recalling the total probability theorem, the probability of correct decision is given by P[C] D P[a0 D a0 ] O D
M X nD1
M X nD1
M XZ nD1 Rn
Given a desired signal corrupted by noise, in general the notion of sufcient statistics applies to any signal, or sequence of samples, that allows the optimum detection of the desired signal. In other words, no information is lost in considering a set of sufcient statistics instead of the received signal. A particular case is represented by transformations that allow reconstruction of a signal using the basis identied by the desired signal. For example, considering the basis fe j2 f t ; t 2 <; f 2 Bg to represent a real-valued signal with passband B in the presence of additive noise, that is the Fourier transform of the noisy signal ltered by an ideal lter with passband B, we are able to reconstruct the noisy signal within the passband of the desired signal; therefore the noisy signal ltered by a lter with passband B is a sufcient statistic. 2 Here we use the formulation (1.377) for real-valued signals; we would get the same results using the formulation (1.380) for complex-valued signals.
441
We dene the indicator function of the set Rn as 2 Rn In D 1 0 elsewhere Then (6.18) becomes Z P[C] D
M X < I nD1
(6.19)
In pn prja0 . j n/ d
(6.20)
The integrand function consists of M terms but, being the M regions non-overlapping, for each value of only one of the terms is different from zero. Therefore the maximum value of the integrand function for each value of , and hence of the integral, is achieved if for each value of we select among M terms the term that yields the maximum value of pn prja0 . j n/. Thus we have the following decision criterion.3 Maximum a posteriori probability (MAP) criterion: 2 Rm Using the Bayes rule prja0 . j n/ D pr ./ the decision criterion becomes a0 D arg max P[a0 D n j r D ] O
n
a0 D m O
(6.22)
P[a0 D n j r D ] pn
(6.23)
(6.24)
In other words, given that we observe , the signal detected by (6.24) has the largest probability of having been transmitted. The probabilities P[a0 D n j r D ]; n D 1; : : : ; M, are the a posteriori probabilities. We give a simple example of application of the MAP criterion for I D 1 and M D 3. Let the function pn pr ja0 . j n/, n D 1; 2; 3, be given as shown in Figure 6.2. If we indicate with 1 , 2 , and 3 the intersection points of the various functions as illustrated in Figure 6.2, it is easy to verify that R1 D . 1; 1 ] R2 D .1 ; 2 ] [ .3 ; C1/ R3 D .2 ; 3 ] (6.25)
arg means argument; for a function f .x; n/, m D arg max f .x; n/
n
(6.21)
denotes the value of m that coincides with the value of n for which the function f .x; n/ is maximum for a given x. If two or more values of n that maximize f .x; n/ exist, a random choice is made to determine m. For a complex number c, arg.c/ denotes the phase of c.
442
p1 pr|a ( |1)
0
p3 pr|a ( |3)
0
p2 pr|a ( |2)
0
Maximum likelihood (ML) criterion. If the signals are equally likely a priori, i.e. pn D 1=M, 8n, the criterion (6.22) becomes 2 Rm a0 D m O if m D arg max prja0 . j n/
n
(6.26)
The ML criterion leads to choosing that value of n for which the conditional probability that r D is observed given a0 D n is maximum. In some texts the ML criterion is formulated via the denition of the likelihood ratios: Ln ./ D prja0 . j n/ prja0 . j 1/ n D 1; 2; : : : ; M (6.27)
(6.28)
1 jj N0
sn jj
(6.29)
sn jj2
(6.30)
Hence the ML criterion coincides with the minimum distance criterion: decide for the signal vector sm , which is closest to the received signal vector . Moreover, the decision regions fRn g, n D 1; : : : ; M, are easily determined. An example is given in Figure 6.3 for the three signals of Example 1.2.2 on page 10. Considering a pair of vectors si ; s j , we draw the straight line of points that are equidistant from si and s j : this straight line denes the boundary between Ri and R j . The procedure then is repeated for every pair of vectors. The decision region associated with each vector sn is given by the intersection of two half-planes as illustrated in Figure 6.3.
Theorem of irrelevance
With regard to the decision process, we introduce a theorem that formalizes the distinction previously mentioned between relevant and irrelevant components of the received signal.
443
2 A T 2 R3 s1 0 R1 A T 2 1 s3 s 2 R2
Figure 6.3. Construction of decision regions for the constellation of the Example 1.2.2.
Let us assume that the signal vector r can be split into two parts, r D [r1 ; r2 ]. Then, under the hypothesis a0 D n, prja0 . j n/ D pr1 ;r2 ja0 . 1 ; 2 j n/ which, recalling the denition of conditional probability, can be rewritten as pr1 ;r2 ja0 . 1 ; 2 j n/ D pr2 jr1 ;a0 . 2 j 1 ; n/ pr1 ja0 . 1 j n/ Substitution of (6.32) into (6.22) leads to the following result. Theorem 6.1 If pr2 jr1 ;a0 . 2 j 1 ; n/ does not depend on the particular value n assumed by a0 , that is if pr2 jr1 ;a0 . 2 j 1 ; n/ D pr2 jr1 . 2 j 1 / (6.33) (6.32) (6.31)
then the optimum receiver can disregard the component r2 and base its decision only on the component r1 . Corollary 6.1 A sufcient condition to disregard r2 is that pr2 jr1 ;a0 . 2 j 1 ; n/ D pr2 . 2 / We illustrate the theorem by the following examples. Example 6.1.1 The system (6.2) is represented using a larger basis, as illustrated in Figure 6.4, where the noise (6.5) has two components w1 D w w2 D w? (6.35) (6.34)
444
s1 s
2
w1 r = w +s 1 1 n r =w 2 2 w2
We note that the received signal vector r2 coincides with the noise vector w2 that is statistically independent of w1 and sn . Therefore we have pr2 jr1 ;a0 . 2 j 1 ; n/ D pr2 . 2 / (6.36)
hence, by Corollary 6.1, the component r2 D w? can be disregarded by the optimum receiver. Example 6.1.2 In the system shown in Figure 6.5, the noise vectors w1 and w2 are statistically independent. As r2 D r1 C w2 , if r1 is known, then r2 depends only on the noise w2 , that is independent of the particular sn transmitted. Then pr2 jr1 ;a0 . 2 j 1 ; n/ D pw2 . 2 1/ (6.37)
does not depend on n: therefore (6.33) is veried and r2 is irrelevant. Example 6.1.3 As in the previous example, the noise vectors w1 and w2 in Figure 6.6 are statistically independent. Under this condition, however, r2 cannot be disregarded by the optimum receiver: in fact, from r2 D w2 C w1 and w1 D r1 sn , we get pr2 jr1 ;a0 . 2 j 1 ; n/ D pw2 . 2 that depends explicitly on n.
s1 s s
2
w1 / D pw2 . 2
1 C sn /
(6.38)
w1
r = w + r = w + w + sn 2 2 1 2 1
r = w +s 1 1 n
Figure 6.5. Example 6.1.2: the vector r2 is irrelevant.
445
w1 s1 s s
2
r = w2 + w 1 2 r = w +s 1 1 n
w1 s1 s
2
r =w 2 2 r = r +r 2 1 r = w +s 1 1 n
Example 6.1.4 In Figure 6.7, if the noise vectors w1 and w2 are statistically independent, observing (6.34) the signal r2 can be neglected by the optimum receiver. In fact, it is: pr2 jr1 ;a0 . 2 j 1 ; n/ D pw2 jw1 ;a0 . 2 j 1 sn ; n/ D pw2 . 2 / (6.39)
that does not depend on n. We note that Example 6.1.1 is a particular case of Example 6.1.4.
We note that the lter on branch i has impulse response given by rect..t yields the output Z t yi .t/ D r. / i . / d
t t0
(6.41)
which sampled at t D t0 , from (6.14), yields yi .t0 / D ri . An equivalent implementation is based on the equivalence illustrated in Figure 6.9, where a correlation demodulator is substituted by a matched lter with impulse response i .t0 t/.
446
*(t)
i
t0 t (.) d t-t 0
(a)
r(t)
ri
t0 r(t) * (t -t) 0
i
ri
(b)
Figure 6.9. (a) Correlation demodulator and equivalent (b) matched lter (MF) demodulator.
Implementation type 2. Using (1.39), from jj the ML criterion becomes a0 D arg max Reh; sn i O
n
2Re[h; sn i]
(6.42)
jjsn jj2 2
(6.43) En 2 (6.44)
D arg max Re
n
C1 1
.t/sn .t/ dt
jsn .t/j2 dt
(6.45)
The implementation of (6.44) is illustrated in Figure 6.10, whereas the equivalent criterion (6.43) is also given in Figure 6.8.
447
E1 2 E2 2
U1
U2
^ a0=arg max Un n ^ a0
EM 2
UM
Typically, in the applications we have I < M, hence implementation type 1 is more convenient.
Error probability
In general, the error probability of the system is dened as O Pe D P[E] D P[a0 6D a0 ] D 1 P[C] (6.46)
where P[C] is given by (6.18). Using the total probability theorem, we express the error probability as
M X nD1
Pe D
pn P[r 2 Rn j a0 D n] =
(6.47)
We examine the case of two signals, whose vector representation is illustrated in Figure 6.11. For convenience we choose 2 parallel to the line joining s1 and s2 . Independently of the basis system, assuming real, from (1.41) the squared distance between the two signals is given by p (6.48) d 2 D E 1 C E 2 2 E 1 E 2 where Z Ei D
1 C1
jsi .t/j2 dt
i D 1; 2
(6.49)
448
and
s1 .t/s2 .t/ dt D p E1 E2 In the case of two equally likely signals, equation (6.47) becomes 1
C1
(6.50)
(6.51)
We assume that s2 is transmitted, which means a0 D 2. Given the received signal vector r as in Figure 6.11, we get a decision error if the noise w D r s2 has a projection on the line joining s1 and s2 that is smaller than d=2. As for all projections on an orthonormal basis, the noise component w2 is Gaussian with zero mean and variance N0 2 Then the conditional error probability is given by P[r 2 R2 j a0 D 2] D P w2 < = I2 D where Q.a/ D Z (6.52) d d DQ 2 2 I
(6.53)
b2 1 (6.54) p e 2 db 2 a is the Gaussian complementary distribution function whose values are reported in Appendix 6.A. Likewise, it is d d (6.55) DQ P[r 2 R1 j a0 D 1] D P w2 > = 2 2 I
C1
d 2 I
(6.56)
449
Observation 6.1 The error probability depends only on the ratio between the distance of the signals of the constellation at the decision point and the standard deviation per dimension of the noise. For this reason it is useful to dene the following signal-to-noise ratio at the decision point D d 2 I 2 (6.57)
We will now derive an alternative expression for Pe as a function of the modulator parameters. Substitution of (6.48) and (6.52) in (6.56) yields 0s 1 p E 1 C E 2 2 E 1 E 2 A (6.58) Pe D Q @ 2N0 If E 1 D E 2 D E s , we get s Pe D Q E s .1 / N0 ! (6.59)
6.1.1
We let M D 2 and E 1 D E 2 D E s .
Antipodal signals ( = 1)
The signal set is composed of two antipodal signals: s1 .t/ D s.t/ For E s D RT
0
s2 .t/ D
s.t/
(6.60)
Es
(6.61)
The basis has only one element, I D 1, with s.t/ .t/ D p Es The vector representation of Figure 6.12 follows, where p p s2 D s1 D Es Es The implementation type 1 of the optimum ML receiver is depicted in Figure 6.13. (6.62)
(6.63)
450
s2 - E
s
s1 0 d=2 Es Es
r(t) * (T-t)
r>0 , ^ =1 a r<0
0 , ^ =2 a 0
^ a0
As D
A modulation technique with antipodal signals is binary phase shift keying (2-PSK or BPSK), where s.t/, dened by (6.60), is shown in Figure 6.14. In this case s1 .t/ D A cos.2 f 0 t C '0 / s2 .t/ D A cos.2 f 0 t C '0 C / D s1 .t/ 0<t <T 0<t <T (6.65) (6.66)
Orthogonal signals ( = 0)
We consider the two signals s1 .t/ D A cos.2 f 0 t/ s2 .t/ D A sin.2 f 0 t/ 0<t <T (6.67)
Observing (1.71), if f 0 D k=2T , k integer, or else f 0 1=T , then Es D E1 D E2 D A2 T 2 and ' 0 (6.68)
A basis is composed of the signals themselves r 2 s1 .t/ cos.2 f 0 t/ D 1 .t/ D p T Es and s2 .t/ D 2 .t/ D p Es We note that
1 .t/
0<t <T
(6.69)
2 sin .2 f 0 t/ T
0<t <T
(6.70)
and
2 .t/
451
A/2
s(t)
A/2
0
t
Figure 6.14. Plot of s.t/ D A cos.2 f0 t C '0 /, 0 < t < T, for f0 D 2=T and '0 D
=2.
2 Es s2 2E s
s1 0 Es 1
The vector representation is given in Figure 6.15. As the two vectors s1 and s2 are p p orthogonal, their distance is 2Es . This distance is reduced by a factor of 2 as compared to the case of antipodal signals with the same value of Es . The optimum ML receiver is depicted in Figure 6.16. As D 0, (6.59) becomes s ! Es Pe D Q (6.71) N0
452
T s* (T-t) r(t)
1
U >U , ^ =1 a
1 2 0
^ a0
T s* (T-t)
2
U <U , ^ =2 a
1 2 0
Figure 6.17. Error probability as a function of Es =N0 for binary antipodal and orthogonal signalling.
From the curves of probability of error versus E s =N0 plotted in Figure 6.17, we note that for a given Pe we have a loss of 3 dB in E s =N0 for the orthogonal signalling scheme as compared to the antipodal scheme. We examine in detail another binary orthogonal signalling scheme.
Binary FSK
We consider the two signals of Figure 6.18, given by s1 .t/ D A cos.2. f 0 f d /t C '0 / s2 .t/ D A cos.2. f 0 C f d /t C '0 / 0<t <T (6.72)
453
s (t)
0
t
s (t)
0
t
Figure 6.18. Binary FSK signals with f0 D 2=T, fd D 0:3=T and '0 D 0.
We have two windowed sinusoidal functions, one with frequency f 0 f d and the other with frequency f 0 C f d ; f 0 is called the carrier frequency, and f d is the frequency deviation. Also in this case, if f 0 f d 1=T , it holds Es D E1 D E2 ' and Z D
C1 1
A2 T 2
(6.73)
E s .1 / N0
! (6.75)
Introducing the modulation index h as the ratio between the frequency deviation f d and the Nyquist frequency of the transmission system equal to 1=.2T /, we have hD Therefore D sinc.2h/. fd D 2 fd T 1=.2T / (6.76)
454
0.8
0.6
0.4
0.2
0.2
0.4
0.5
1.5
2 h
2.5
3.5
From the plot of as a function of h illustrated in Figure 6.19, we get the minimum p value of , min D 0:22, for h D 0:715 and Pe D Q. 1:22E s =N0 /, with a gain of 0.7 dB in E s =N0 as compared to the case D 0, and a loss of 2.3 dB with respect to antipodal signalling. From (6.74) we have D 0 for h D 1=2, that is for f d D 1=.4T /: in this case we speak of minimum shift keying (MSK).4 There are other values of h (1; 1:5; 2; : : : ) that yield D 0, however, they imply larger f d , with the consequent requirement of larger channel bandwidth.
6.1.2
Let fsn .t/g, n D 1; : : : ; M, be M equally likely signals, with squared distances Z C1 2 2 jsn .t/ sm .t/j2 dt d .sn ; sm / D dnm D
1
(6.77)
and
2 2 dmin D min dnm n;m
(6.78)
Upper bound
We assume sm is transmitted. An error event that leads to choosing sn is expressed as Enm D fr : d.r; sn / < d.r; sm /g (6.79)
In fact, MSK also requires that the phase of the modulated signal be continuous (see Section 18.5).
455
Using (6.53), the probability of the event Enm is given by dnm P[Enm ] D Q 2 I
M X mD1
(6.80)
In general, from (6.47), the error probability in the case of equally likely signals is Pe D P[E j sm ] 1 M (6.81)
On the other hand, as sm is transmitted, an error occurs if sn is chosen, n D 1; : : : ; M, n 6D m. Therefore, applying the union bound, we have " # M M [ X Enm P[Enm ] (6.82) P[E j sm ] D P
nD1;n6Dm nD1;n6Dm
Since dmin dnm , then Q.dnm =.2 I // Q.dmin =.2 I //, and a looser bound than (6.83) is given by dmin Pe .M 1/Q (6.84) 2 I
Lower bound
Given sn , let dmin;n be the distance of sn from the nearest signal. Limiting the evaluation of the error probability to error events that are associated to the nearest signals, we obtain the following lower bound M 1 X dmin;n (6.85) Q Pe M nD1 2 I We get a looser bound by introducing Nmin , the number of signals fsn .t/g whose distance is dmin from the nearest signal: 2 Nmin M. Limiting the equation in (6.85) to such signals, we have dmin Nmin Pe Q (6.86) M 2 I In other words, given sm , an error event E is reduced to considering only a signal at minimum distance, if there is one. In the particular case of dmin;n D dmin , we have Nmin D M, and Pe Q.dmin =.2 I //. For example, for the constellation of Figure 6.3 with M D 3 we have r T Ap Ap d23 D T d13 D A T (6.87) d12 D 2 2 2 p Then dmin D A T =2 and Nmin D 3.
456
6.2
With reference to Figure 6.20, we discuss now some aspects of a communication system, where the information message consists of a sequence of information bits fb` g generated at instants `Tb . Some bits may be inserted in the message fb` g to generate an encoded message fcm g, according to rules that will be investigated in Chapters 11 and 12. The bits of fcm g are then mapped to the symbols fak g, which assume values in an M-ary alphabet and are generated at instants kT .5 The value of the generic symbol ak is then modulated, that is, it is associated with a waveform, according to the scheme of Figure 6.1: ak ! sak .t Therefore the transmitter generates the signal s.t/ D
C1 X kD 1
kT /
(6.88)
sak .t
kT /
(6.89)
which is sent over the transmission channel. Let sCh .t/ be the signal at the output of the transmission channel, which is assumed to introduce additive white Gaussian noise w.t/ with PSD N0 =2. Note that the system of Figure 6.1 has been investigated assuming that an isolated waveform is transmitted. In the model of Figure 6.20, the transmission of a waveform is repeated every symbol period T . However, with reference to Figure 6.8, assuming that the transmitted waveforms do not give rise to intersymbol interference (ISI) at the demodulator
Note that there are systems in which encoder and modulator are jointly considered, see, for example, Chapter 12. In that case the notion of binary channel cannot be referred to the transmission of the sequence fcm g.
6.2. Simplied model of a transmission system and denition of binary channel 457
output6 we can still study the system assuming that an isolated symbol is transmitted, for example, the symbol a0 transmitted at instant t D 0. At the receiver, the bits fc` g are obtained by inverse bit mapping from the detected Q O message fak g. The information bits fb` g are then recovered by a decoding process. O Denition 6.1 The transformation that maps cm into cm is called a binary channel. It is characterized by Q the bit rate 1=Tcod , which is the transmission rate of the bits of the sequence fcm g, and by the bit error probability Q Pbit D PBC D P[cm 6D cm ] cm ; cm 2 f0; 1g Q (6.90)
In the case of a binary symmetric channel (BSC), it is assumed that P[cm 6D cm j cm D Q 0] D P[cm 6D cm j cm D 1]. We say that the BSC is memoryless if, for every choice of N Q distinct instants m 1 ; m 2 ; : : : ; m N , the following relation holds: Q Q P[cm 1 6D cm 1 ; cm 2 6D cm 2 ; : : : ; cm N 6D cm N ] Q Q Q D P[cm 1 6D cm 1 ] P[cm 2 6D cm 2 ] : : : P[cm N 6D cm N ] Q (6.91)
In a memoryless binary symmetric channel the probability distribution of fcm g is obtained Q from that of fcm g and PBC according to the statistical model shown in Figure 6.21. We note that the aim of the channel encoder is to introduce redundancy in the sequence fcm g, which is exploited by the decoder to detect and/or correct errors introduced by the binary channel. The overall objective of the transmission system is to reproduce the sequence of information bits fb` g with a high degree of reliability, measured by the bit error probability
.dec/ O Pbit D P[b` 6D b` ]
(6.92)
Absence of ISI in this context means that the optimum reception of the waveform transmitted at instant kT , sak .t kT /, is not inuenced by the presence of the waveforms associated with symbols transmitted at other instants. For example, all signalling schemes that employ pulses with nite duration in the interval .0; T / do not give rise to ISI. However, this is a particular case of the Nyquist criterion for the absence of ISI that will be discussed in Section 7.3.3.
458
.dec/ Typically, it is required Pbit ' 10 2 10 3 for PCM or ADPCM coded speech (see .dec/ Chapter 5) and Pbit ' 10 7 10 11 for data messages.
6.2. Simplied model of a transmission system and denition of binary channel 459
2. Number of information bits per modulation interval : via the channel encoder (COD) and the bit-mapper (BMAP), L b information bits of the message fb` g are mapped in an M-ary symbol, ak . In general we have L b log2 M (6.94)
where the equality holds for a system without coding, or, with abuse of language, for an uncoded system. In this case we also have RI D 3. Symbol period : T D Tb L b 4. Statistical power of the desired signal at the receiver input: MsCh D E sCh T (6.97) (6.96) log2 M I (6.95)
We note that, for continuous transmission (see Chapter 7), MsCh is nite and consequently we dene E sCh D MsCh T ; on the other hand, for transmission of an isolated pulse E sCh is nite and we dene MsCh through (6.97). 5. Average energy per dimension: EI D 6. Average energy per information bit: Eb D For an uncoded system, (6.99) becomes: Eb D E sCh log2 M (6.100) EI Es D Ch RI Lb (6.99) E sCh I (6.98)
7. Conventional minimum bandwidth of the modulated signal : Bmin D 1 2T 1 T for baseband signals for passband signals (6.101) (6.102)
Bmin D
460
For the orthogonal and biorthogonal signals of Section 6.7, the denition of Bmin will be different and will include the factor 1=M. 8. Spectral efciency: D Lb 1=Tb D Bmin Bmin T (6.103)
In practice measures how many bits per unit of time are sent over a channel with the conventional bandwidth Bmin . In terms of R I , from (6.93), we have D RI I Bmin T (6.104)
Later we will see that, for most uncoded systems, R I D =2. 9. Conventional signal-to-noise ratio at the receiver input: 0D MsCh E sCh D .N0 =2/2Bmin N0 Bmin T (6.105)
In general 0 expresses the ratio between the statistical power of the desired signal at the receiver input and the statistical power of the noise measured with respect to the conventional bandwidth Bmin . We note that, for the same value of N0 =2, if Bmin doubles, the statistical power must also double to maintain a given ratio 0. 10. Signal-to-noise ratio per dimension: 0I D 2E sCh EI D N0 =2 N0 I (6.106)
is the ratio between the energy per dimension of an isolated pulse E I and the noise variance per dimension I2 given by (6.13). Using (6.99), the general relation becomes 0 I D 2R I Eb N0 (6.107)
It is interesting to observe that in most modulation systems it turns out 0 I D 0. 11. Link budget: if the receiver is matched to the transmission medium for the maximum transfer of power, from (4.92) we obtain an alternative expression of (6.105) given by 0D PsCh kTwi Bmin (6.108)
We observe that (6.105) is useful to analyze the system, and (6.108) is usually employed to evaluate the link budget. In the next sections some examples of modulation systems without channel coding are illustrated.
461
6.3
Pulse amplitude modulation, also called amplitude shift keying (ASK), is the rst example of multilevel baseband signalling, i.e. M may take values larger than 2. Let h Tx be a realvalued nite-energy pulse with support .0; t0 /; a transmitted isolated pulse is expressed as sn .t/ D n h Tx .t/ where n D 2n 1 M (6.110) t 2< n D 1; 2; : : : ; M (6.109)
In other words, PAM signals are obtained by modulating in amplitude the pulse shape h Tx . Energy of sn :
2 E n D n E h
Z Eh D
C1 1
jh Tx .t/j2 dt
(6.111)
(6.112)
(6.115)
The transmitter is shown in Figure 6.23. The bit mapper is composed of a serial-to-parallel (S/P) converter followed by a map that translates a sequence of log2 M bits into the corresponding value of a0 . The map is a Gray encoder (see Appendix 6.B). An example for M D 8 is illustrated in Table 6.1. The symbol n is input to an interpolator lter with impulse response h Tx . The lter output yields the transmitted signal sa0 .
7
iD
M.M C 1/ 2
M X iD1
i2 D
M.M C 1/.2M C 1/ 6
M X iD1
i3 D
M C1 2 2
462
Gray coding of symbols (M D 8) Three-bit sequence 000 001 011 010 110 111 101 100 n 7 5 3 1 1 3 5 7 a0 1 2 3 4 5 6 7 8
The type 1 implementation of the ML receiver is shown in Figure 6.24 and consists of a matched lter to h Tx followed by a sampler. In this case, from (6.15) r is given by r D sn C w (6.116) where sn D n .d=2/, n D 1; : : : ; M, and w is a real-valued Gaussian r.v. with zero mean and variance N0 =2. From the observation of r, a threshold detector yields the detected symbol a0 . The O transmitted bits are then recovered by an inverse bit mapper. Minimum bandwidth of the modulated signal, equal to the Nyquist frequency, see Denition 7.1 on page 559: 1 (6.117) Bmin D 2T
463
Figure 6.24. ML receiver, implementation type 1, of a PAM system for an isolated pulse. The p thresholds are set at .2n M/ Eh D .n .M=2// d, n D 1; 2; : : : ; M 1.
Spectral efciency: D Signal-to-noise ratio: .1=T / log2 M D 2 log2 M (bit/s/Hz) 1=.2T / (6.118)
Note that (6.119) expresses 0 as the ratio between the signal energy and the variance of the noise component: therefore, as I D 1, it follows that 0 I D 0. Symbol error probability: from the total probability theorem, letting d D 2 considering the outer constellation symbols separately from the others, we have P[E j s M ] D P[E j s1 ] D P[a0 6D 1 j a0 D 1] O DP r> 1 M 2 d j a0 D 1 M 2 d j a0 D 1 (6.120) M d 2 p E h , and
d D P 1 C w > 1 2 D P .1
d M/ C w > 1 2
d DP w> 2 DQ d 2 I
464
and O n D 2; : : : ; M 1 P[E j sn ] D P[a0 6D n j a0 D n] d d [ r > .n C 1/ j a0 D n D P r < .n 1/ 2 2 d d d d C P n C w > .n C 1/ D P n C w < .n 1/ 2 2 2 2 d d DP w< CP w> 2 2 d D 2Q 2 I where I2 D N0 =2. Then, for equally likely symbols we have d d 1 C .M 2/2Q 2Q Pe D M 2 I 2 I 1 d D2 1 Q M 2 I In terms of E s we get8 Pe D 2 1 1 M Q s ! (6.123)
(6.121)
(6.122)
6 M2
Es 1 N0
(6.124)
Assuming that Gray coding is adopted at the transmitter, the bit error probability is given by Pbit ' Pe log2 M valid for 0 1 (6.125)
Equation (6.125) expresses the fact that, for 0 sufciently large, if an error event occurs, it is very likely that one of the symbols at the minimum distance from the transmitted symbol is detected. Thus with high probability only one bit of the log2 M bits associated with the transmitted symbol is incorrectly recovered. Curves of Pbit as a function of 0 are shown in Figure 6.25 for different values of M. In the Appendix 6.C, two other baseband modulation schemes are described: pulse position modulation (PPM) and pulse duration modulation (PDM).
8
These results are valid also for continuous transmission with modulation period T , assuming absence of ISI at the decision point; in other words, the autocorrelation function of h Tx .t/ must be a Nyquist pulse (see Section 7.3.3).
465
10
10
M=16
10
M=8
Pbit
10
M=4
10
M=2
10
10
15 20 =2E /N (dB) s 0
25
30
35
6.4
Phase-shift keying is an example of passband modulation. Let h Tx be a real-valued niteenergy baseband pulse with support .0; t0 /. Let9 'n D .2n M 1/ n D 1; : : : ; M (6.126)
then the generic transmitted pulse is given by sn .t/ D h Tx .t/ cos.2 f 0 t C 'n / t 2< n D 1; : : : ; M (6.127)
that is, signals are obtained by choosing one of the M possible values of the phase of a sinusoidal function with frequency f 0 , modulated by h Tx .10 In the following sections we will denote by k the r.v. that determines the transmitted signal phase at instant kT . Consequently, the values of k are given by 'n , n D 1; : : : ; M. In this section we consider the case of an isolated pulse transmitted at instant k D 0. An alternative expression of (6.127) is given by sn .t/ D Re[e j'n h Tx .t/e j2 f 0 t ] (6.128)
A more general denition of 'n is given by 'n D M .2n 1/ C '0 , where '0 is a constant phase. 10 We obtain (6.66) by assuming h .t/ D w .t/, where w is the rectangular window of duration T dened Tx T T 9
in (1.473).
466
Moreover, setting n D e j'n D e j M .2n we have sn .t/ D n;I h Tx .t/ cos.2 f 0 t/ where n;I D Re[n ] D cos n;Q .2n M D Im[n ] D sin .2n M 1/ 1/ (6.131) (6.132) n;Q h Tx .t/ sin.2 f 0 t/ (6.130)
1/
(6.129)
Energy of sn : if f 0 is greater than the bandwidth of h Tx , using Parseval theorem we get En D Average energy of the system: Es D Basis functions: s
1 .t/ M 1 X Eh En D M nD1 2
Eh 2
(6.133)
(6.134)
2 .t/
(6.135)
(6.136)
1/ ; sin .2n M
iT 1/
n D 1; 2; : : : ; M
(6.137)
q We note that the desired signal at the decision point, s n , aside from the factor E h , 2 coincides with n . Note that the signal constellation lies on a circle and the various vectors differ in the phase 'n . The minimum distance is equal to p p D 2E h sin (6.138) dmin D 2 E s sin M M
467
In Figure 6.26, the projections of sn on the axis 1 (in phase) and on the axis 2 (quadrature) are also represented, together with the Gray coding of the various symbols represented by the bits b1 ; b2 ; b3 . A PSK transmitter for M D 8 is shown in Figure 6.27. The bit mapper maps a sequence of log2 M bits to a constellation point with value n . The quadrature components n;I and n;Q are input to interpolator lters h Tx . The lter output signals are multiplied by the carrier signal, cos.2 f 0 t/, and by the carrier signal phase-shifted by =2, for example, by a Hilbert lter, sin.2 f 0 t/, respectively. The transmitted signal (6.130) is obtained by adding the two components. The type 1 implementation of the ML receiver is illustrated in Figure 6.28. From the general scheme of Figure 6.8, we note that the basis functions (6.136) are implemented partially by a correlator with a sinusoidal signal, and partially by a lter matched to h Tx . From Figure 6.26 we note that the decision regions are angular sectors with phase 2=M. For M D 2; 4; 8, simple decision rules can be dened. For M > 8 detection can be made by observing the phase vr of the received vector11 r D [r I ; r Q ]T .
11 For the sake of notation uniformity with the following chapters, the components r and r of r will be indicated 1 2 by r I and r Q , respectively.
468
Figure 6.28. ML receiver, implementation type 1, of an M-PSK system for an isolated pulse. Thresholds are set at .2 =M/n, n D 1; : : : ; M.
Minimum bandwidth of the modulated signal (passband signal): Bmin D Spectral efciency: D .1=T / log2 M D log2 M (bit/s/Hz) 1=T (6.140) 1 T (6.139)
469
Signal-to-noise ratio:
We note that 0 also expresses the ratio between the energy per dimension and the variance of the noise components; moreover 0 I D 0 if M > 2. Symbol error probability: with equally likely signals, exploiting the symmetry of the signalling scheme we get Pe D P[E j sn ] D1 D1 D1 P[C j sn ] P[r 2 Rn j a0 D n] ZZ pr . I ; Q j a0 D n/ d I d Q
Rn
(6.142)
where the angular sector Rn is illustrated in Figure 6.29. For a0 D n we get r D w C sn , where sn is given by (6.137); then, observing (6.16), (6.142) becomes ZZ 2 2 p p 1 1 I exp E s cos 'n C Q E s sin 'n Pe D 1 d I d Q N0 Rn N 0 (6.143) Using polar coordinates, we get Z Pe D 1
M M
p .z/ dz
(6.144)
w sm
2 M
vr
470
where p .z/ D e
E s =N0
( 1C
s Q
2E s cos z N0
!#) (6.145)
for z . The integral (6.144) cannot be solved in closed form. If E s =N0 1, for M 4 we can use the approximation (6.363) in (6.145) to obtain s Es Es sin2 z p .z/ ' cos ze N0 (6.146) N0 In turn, substituting (6.146) in (6.144), and observing (6.141), we get s ! 2E s Pe D 2Q sin N0 M p D 2Q 20 sin M
(6.147)
Assuming that Gray coding is adopted at the transmitter, the bit error probability is given by Pbit D Pe log2 M valid for Es 1 N0 (6.148)
Curves of Pbit as a function of 0 D E s =N0 are shown in Figure 6.30. We consider in detail two particular cases.
Moreover, it is D 1. This result is due to the fact that a BPSK system does not efciently use the available bandwidth 1=T : in fact only half of the band carries information. The information in the other half can be deduced by symmetry and is therefore redundant.
471
10
10
M=32
10
M=16
Pbit
10
M=8
10
M=4
M=2
10
10
15 20 =E /N (dB) s 0
25
30
35
From (6.64), obtained for antipodal signals, and using (6.141), the evaluation of Pe yields Pe D Pbit s DQ 2E s N0 ! (6.151)
p D Q. 20/ The transmitter and the receiver for a BPSK system are shown in Figure 6.32 and have a very simple implementation. The bit mapper of the transmitter maps 0 in 1 and 1 in C1 to generate NRZ binary data (see Appendix 7.A). At the receiver, the decision element implements the sign function to detect NRZ binary data. The inverse bit mapping to recover the bits of the information message is straightforward.
472
Figure 6.32. Schemes of transmitter and receiver for a BPSK system with '0 D 0.
1 2Q
(6.153)
(6.154)
The QPSK transmitter is obtained by simplication of the general scheme (6.130), as illustrated in Figure 6.34. The binary bit maps are given in Table 6.2. The ML receiver for QPSK is illustrated in Figure 6.35. As the decision thresholds are set at .0; =2; ; 3=2/
473
2 b 1 b2 01 11
s2
s1
E s = Eh /2
1 s3 00 s4 10
474
(see Figure 6.33), decisions can be made independently on r I and r Q , using a simple threshold detector with threshold set at zero. We observe that, for h Tx .t/ D K wT .t/, the transmitter lter is a simple holder. At the receiver the matched lter plus sampler becomes an integrator that is cleared before each integration over a symbol period of duration T . In other words, it consists of an integrateand-dump.
6.5
We assume now that the receiver recovers the carrier signal, except for a phase offset of 'a . In particular, with reference to the scheme p Figure 6.28, the reconstructed carrier is cos.2 f 0 t 'a /. In this case s n coincides with E s e j'a n , where n is given by (6.129). Consequently, it is as if the constellation at the receiver were rotated by 'a . To prevent this problem there are two strategies. By the coherent method, a receiver estimates 'a from the received signal, and considers the original constellation for detection, using the signal O O r e j 'a , where 'a is the estimate of 'a . By the differential non-coherent method, a receiver detects the data using the difference between the phases of signals at successive sampling instants. In other words for M-PSK, the phase of the transmitted signal at instant kT is given by (6.126), with k 2 .2M 1/ 3 ; ;:::; M M M (6.156)
475
(6.157)
that is, the phase associated with the transmitted signal at instant kT is equal to that transmitted at the previous instant .k 1/T plus the increment k , which can assume one of M values. We note that the decision thresholds for k are now placed at .=M/.2n 1/, n D 1; : : : ; M. For a phase offset equal to 'a introduced by the channel, the phase of the signal at the detection point becomes
k
0 k
C 'a
(6.158)
In any case,
k k 1
D k
(6.159)
and the ambiguity of 'a is removed. For phase-modulated signals, three differential noncoherent receivers that determine an estimate of (6.159) are discussed in Chapter 18.
6.5.1
For E s =N0 1, using the denition of the Marcum function Q 1 .; / (see Appendix 6.A) it can be shown that the error probability of an isolated symbol is approximated by the following bound [2, 3] s s ! Es Es ; 1 sin 1 C sin Pe 1 C Q 1 N0 M N0 M (6.160) s s ! Es Es Q1 1 C sin 1 sin ; N0 M N0 M Moreover, if M is large, the approximation (6.369) can be used and we get "s r r # Es 1 C sin 1 sin Pe ' 2Q N0 M M s ' 2Q Es sin N0 M ! (6.161)
12 Note that we consider a differential non-coherent receiver with which is associated a differential symbol
encoder at the transmitter (see (6.157)) or ((6.169)). However, as we will see in the next section, a differential encoder and a coherent receiver can be used.
476
For Gray coding of the values of k in (6.156), the bit error probability is given by Pbit D Pe log2 M (6.162)
where Pe is given by (6.161). For M D 2, the exact formula of the error probability is [2, 3] Pbit D Pe D 1 e 2 For M D 4, the exact formula is [2, 3] Pe D 2Q 1 .a; b/ where s aD Es .1 N0 p 1=2/ s bD p Es .1 C 1=2/ N0 (6.165) I0 .ab/ e
0:5.a 2 Cb2 / Es N0
(6.163)
(6.164)
and where the function I0 is dened in (4.216). Using the previous results, a comparison in terms of Pbit between DPSK (6.161) and PSK (6.147) is given in Figure 6.36: we note that, for Pbit D 10 3 , DPSK presents a loss of only 1.2 dB in 0 for M D 2, that increases to 2.3 dB for M D 4, and to 3 dB for M > 4. As a DPSK receiver is simpler as compared to a coherent PSK receiver, in that it does not require recovery of the carrier phase, for M D 2 DPSK is usually preferred to PSK.
10
1
PSK DPSK
2
10
10
Pbit
10
M=2
M=4
M=8
M=16
M=32
10
10
10
15
20 (dB)
25
30
35
477
Note that, if the previously received sample is used as a reference, DPSK gives lower performance with respect to PSK, especially for M 4, because both the current sample and the reference sample are corrupted by noise. This drawback can be mitigated if the reference sample is constructed by using more than one previously received samples [4]. In this way we establish a gradual transition between differential phase demodulation and coherent demodulation. In particular, if the reference sample is constructed using the samples received in the two previous modulation intervals, DPSK and PSK yield similar performance [4].
6.5.2
If 'a is a multiple of 2=M, at the receiver the phase difference can be formed between the phases of two consecutive coherently detected symbols, instead of between the phases of two consecutive samples. In this case, symbols are differentially encoded before modulation.
N where denotes the modulo 2 sum; therefore ck D ck 1 if bk D 0, and13 ck D ck 1 if bk D 1. For the bit map of Table 6.4 we have that bk D 1 causes a phase transition, and bk D 0 causes a phase repetition. Decoder. If fck g are the detected coded bits at the receiver, the information bits are O recovered by O bk D c k . c k O O
1/
D ck ck O O
(6.167)
478
ck Transmitted phase 0 1 0
(rad)
O We note that a phase ambiguity 'a D does not alter the recovered sequence fbk g: in 0 D c 1g and we have O Ok fact, in this case fck g becomes fck O .ck 1/ .ck O O
1
1/ D ck ck O O
O D bk
(6.168)
Multilevel case
Let fdk g be a multilevel information sequence, with dk 2 f0; 1; : : : ; M we have ck D ck
M 1
dk
M
ciated with the bit map is k 2 f=M; 3 =M; : : : ; .2M 1/ =Mg. This encoding and bit-mapping scheme are equivalent to (6.157). At the receiver the information sequence is recovered by O O O (6.170) dk D c k . c k 1 /
M
It is easy to see that an offset equal to j 2 f0; 1; : : : ; .M 1/g in the sequence fck g, O corresponding to a phase offset equal to f0; 2 =M; : : : ; .M 1/2=Mg in f k g, does not O cause errors in fdk g. In fact, O ck j O ck 1 j O D c k . c k 1 / D dk O O (6.171)
M M M M
Performance of a PSK system with differential encoding and coherent demodulation by the scheme of Figure 6.28, is worse as compared to a system with absolute phase encoding. O However, for small Pe , up to values of the order of 0:1, we observe that an error in fck g O causes two errors in fdk g. Approximately, Pe increases by a factor 2,14 which causes a negligible loss in terms of 0. To combine Gray encoding of values of ck with the differential encoding (6.169), a two step procedure is adopted:
14 If we indicate with P e;Ch the channel error probability, then the error probability after decoding is given
(6.172) (6.173)
2 3 8Pe;Ch C 8Pe;Ch
479
values of dk 0 1 2 3 4 5 6 7
1. represent the values of dk with a Gray encoder using a combinatorial table, as illustrated for example in Table 6.5 for M D 8; 2. determine the differentially encoded symbols according to (6.169).
Example 6.5.1 (Differential encoding 2B1Q) We consider a differential encoding scheme for a four-level system that makes the reception insensitive to a possible change of sign of the transmitted sequence. For M D 4 this implies insensitivity to a phase rotation equal to in a 4-PSK signal or to a change of sign in a 4-PAM signal. .1/ .0/ For M D 4 we give the law between the binary representation of dk D .dk ; dk /, .i .1/ .0/ .i dk / 2 f0; 1g, and the binary representation of ck D .ck ; ck /, ck / 2 f0; 1g:
.1/ .1/ .1/ c k D dk c k 1 .0/ .0/ .1/ c k D dk c k
(6.174)
The bit map is given in Table 6.6. The equations of the differential decoder are O.1/ O.1/ O.1/ dk D c k c k 1 O.0/ O.0/ O.1/ dk D c k c k
Table 6.6 Bit map for the differential encoder 2B1Q.
.1/ .0/ ck ck Transmitted symbol ak
(6.175)
0 0 1 1
0 1 0 1
3 1 1 3
480
6.6
Quadrature amplitude modulation is another example of passband modulation. Consider choosing a bit mapper that associates to a sequence of log2 M bits a symbol from a constellation of cardinality M and elements given by the complex numbers n n D 1; 2; : : : ; M (6.176)
If we modulate a symbol of the constellation by a real baseband pulse h Tx with nite energy E h and support .0; t0 /, we obtain the isolated generic transmitted pulse given by sn .t/ D n;I h Tx .t/ cos.2 f 0 t/ n;Q h Tx .t/ sin.2 f 0 t/ t 2< n D 1; : : : ; M (6.177)
where n;I and n;Q denote the real and imaginary part of n , respectively. From (6.176) we also have sn .t/ D Re[n h Tx .t/e j2 f 0 t ] (6.178)
The expression (6.177) indicates that the transmitted signal is obtained by modulating in amplitude two carriers in quadrature. However, if the amplitudes jn j, n D 1; : : : ; M, are not all equal, equation (6.178) suggests that the transmitted signals are obtained not only by varying the phase of the carrier but also the amplitude, hence the name amplitude modulation-phase modulation (AM-PM). In fact QAM may be regarded as an extension of PSK. Energy of sn : if f 0 is larger than the bandwidth of h Tx , we have E n D jn j2 Average energy of the system: Es D
M 1 X En M nD1
Eh 2
(6.179)
(6.180)
For a rectangular constellation M D L 2 , and n I ;I ; n Q ;Q 2 [ .L Then Es D D hence Eh D Es 3 M 1 (6.183) 2 2 .L 3 M 3 1 1/ Eh Eh 2 1/; .L 3/; : : : ; 3; 1; 1; 3; : : : ; .L 1/] (6.181)
(6.182)
481
Basis functions:
basis functions for the signals dened in (6.177) are given by s 2 h Tx .t/ cos.2 f 0 t/ 1 .t/ D Eh s 2 h Tx .t/ sin.2 f 0 t/ 2 .t/ D Eh
(6.184)
as illustrated in Figure 6.37 for various values of M. q We note that, except for the factor E h , in a QAM system s n coincides with n . 2 It is important to observe that for the signals in (6.185) the minimum distance between p two symbols is equal to 2E h , hence p (6.186) dmin D 2E h Consequently, to maintain a given dmin , for every additional bit of information, that is doubling M, we need to increase the average energy of the system by about 3 dB, according to the law 6 2 (6.187) dmin D E s M 1
1 0 (via Q) 11 1 1 11 11 11 00 0 0 00200 00 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 00 0 0 00M=256 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 00 0 0 00 11 11 1 0 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 1 00 11 0 0 00 0M=128 00 1 1 11 11 0 0 00 00 11 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 11 0 0 00 00 1 0 11 1 1 11 11 00 00 0 0 00 00 M=64 00 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 0 0 00 11 11 11 1 1 11 00 00 00 0 0 00 M=32 00 00 1 1 11 1 11 11 0 0 00 0 00 00 1 0 11 11 0 11 11 1 0 1 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 0 11 11 0 0 00 M=16 00 00 1 0 11 11 03 1 00 00 1 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 0 00 00 0 0 00 M=40 00 00 1 0 00 00 01 1 11 11 1 00 00 0 1 11 11 11 1 1 11 00 00 00 0 0 00 11 11 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 11 0 1 11 11 11111111111111 00000000000000 11 1 1 11 0 00 00 00 0 0 00111 11 1 1 11 0 00 00 0 0 00 1 11 11 1 0 11 1 1 11 11 11 00 0 0 00 00 00 1 (via I) 1 1 11 1 11 11 0 0 00 0300 00 1 1 11 11 0 0 00 00 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 11 0 0 00 00 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 11 0 0 00 00 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 11 0 0 00 00 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 11 0 0 00 00 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 11 0 0 00 00 11 1 1 11 11 11 00 0 0 00 00 00 1 1 1 1 11 11 0 0 0 0 00 00 1 1 11 11 0 0 00 00 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 11 0 0 00 00 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 11 0 0 00 00 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 11 0 0 00 00 11 1 1 11 11 11 00 0 0 00 00 00 1 1 1 1 11 11 0 0 0 0 00 00 1 1 11 11 0 0 00 00 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 11 0 0 00 00 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 11 0 0 00 00 11 1 1 11 11 11 00 0 0 00 00 00 1 1 11 1 11 11 0 0 00 0 00 00 1 1 11 11 0 0 00 00 1 0
Figure 6.37. Signal constellations of M-QAM. The term p Eh =2 in (6.185) is normalized to one.
482
s4
1000
s3
1100 3
s2
0100
s1
0000
b 3 b4 00 01 11 10
s8
1001 -3 s12 1011
s7
1101 1 -1 s11 -1 1111
s6
0101 1 s10 0111
s5
0001
s 93
0011
s16
1010
s15 -3
1110
s14
0110
s13
0010
b 1 b2
10
11
01
00
The transmitter of an M-QAM system is illustrated in Figure 6.38 for M D 16. The bit map and the signal constellation of a 16-QAM system are shown in Figure 6.39. We note that the signals that are multiplied by the two carriers are PAM signals: in this example they are 4-PAM signals. The ML receiver for a 16-QAM system is illustrated in Figure 6.40. We note that, as the 16-QAM constellation is rectangular, the decision regions are also rectangular and detection on the I and Q branches can be made independently by observing r I and r Q . In general, however, given r D [r I ; r Q ]T , we need to compute the M distances from the points sn , n D 1; : : : ; M, and choose the nearest to r. The following parameters of QAM systems are equal to those of PSK: 1 Bmin D (6.188) T (6.189) D log2 M
483
Symbol error probability for M D L 2 , rectangular constellation. We rst evaluate the probability p correct decision for a 16-QAM signal. We need to consider the following of cases (d D 2E h ): 2 d n D 1; 4; 13; 16 (6.192) P[C j sn ] D 1 Q 2 I d d 1 2Q n D 2; 3; 5; 8; 9; 12; 14; 15 (6.193) P[C j sn ] D 1 Q 2 I 2 I 2 d P[C j sn ] D 1 2Q n D 6; 7; 10; 11 (6.194) 2 I The probability of error is then given by d d d 2 2:25Q ' 3Q Pe D 1 P[C] D 3Q 2 I 2 I 2 I where the last approximation is valid for large values of d=.2 I /.
(6.195)
484
Another expression can be found in terms of 0 using (6.186), (6.183), and (6.190), Pe ' 4 1 1 p M Q s .M 3 1/ ! 0 (6.197)
Curves of Pbit as a function of 0 are shown in Figure 6.41. We note that, to achieve a given Pbit , if M is increased by a factor 4, we need to increase 0 by 6 dB: in other words, if we increase by one the number of bits per symbol, on average we need an increase of the energy of the system of 3 dB. We arrived at the same result using the notion of dmin in (6.187).
10
1
10
M=256
10
M=64
Pbit
10
M=16
10
M=4
10
10
15 20 =Es/No (dB)
25
30
35
Figure 6.41. Bit error probability as a function of 0 for M-QAM transmission with rectangular constellation.
485
10
PSK M=32
QAM M=256
PSK
QAM M=64
10
Pbit
10
4
10
10
10
15 20 =Es/No (dB)
25
30
35
Figure 6.42. Comparison between PSK and QAM systems in terms of Pbit as a function of 0.
Table 6.7 Gain of QAM with respect to PSK in terms of 0, for given M.
1/
(dB)
2 sin2 .=M/
486
6.7
6.7.1
The isolated generic transmitted pulse, sn , belongs to a set of M orthogonal signals with support .0; t0 / and energy E s , hence Z t0 hsi ; s j i D si .t/s .t/ dt D E s i j i; j D 1; : : : ; M (6.199) j
0
A basis for these signals is simply given by the functions sn .t/ n D 1; : : : ; M (6.200) n .t/ D p Es The vector representations of sets of orthogonal signals for M D 2 and M D 3 are illustrated in Figure 6.43, where p s n D E s [ 0; 0; : : : ; 0; 1; 0; : : : ; 0 ]T (6.201) 1 2 n M p We note that the distance between any two signals is equal to dmin D 2E s . We will now consider a few examples of orthogonal signalling schemes. Example 6.7.1 (Multilevel FSK) 1. Coherent sn .t/ D A sin.2 f n t C '/ where the conditions 1 2T 1 fn C f` D k (k integer) T guarantee orthogonality among the signals. fn fn
1
0<t <T
n D 1; : : : ; M
(6.202)
or else
f1
1 T
(6.203)
2 Es s 2 s1 Es 1
2 Es s 2 s1 Es 3 Es s3 1
(a) M D 2.
(b) M D 3.
Figure 6.43. Vector representations of sets of orthogonal signals for (a) M D 2 and (b) M D 3.
487
0<t <T
n D 1; : : : ; M
(6.204)
(6.205) 1 1 (k integer) or else f 1 fn C f` D k T T guarantee orthogonality among the signals. In (6.204) the uniform r.v. 'n is introduced as each signal has an arbitrary phase. We note that in both cases the bandwidth required by the passband modulation system is proportional to M. In the case of coherent demodulation, we use the denition Bmin D Correspondingly, from (6.103) we have D and from (6.105) 0D As I D M, we have 0I D 0 (6.209) 2E s N0 M (6.208) 2 log2 M M (6.207) M 2T (6.206)
1 T
For non-coherent demodulation, we have Bmin D M=T , D .log2 M/=M, 0 D E s =.N0 M/, and 0 I D 20. Example 6.7.2 (Binary modulation with orthogonal signals) In case we use only two orthogonal signals out of L available signals to realize a binary modulation, for Bmin D L=.2T /, the spectral efciency is 2 L The system is not efcient in exploiting the available bandwidth. D Example 6.7.3 (Code division modulation) Let f pn; j g n D 1; : : : ; L j D 0; : : : ; L 1 (6.211) (6.210)
488
pn; j wTc .t
j Tc /
n D 1; : : : ; M
0 < t < L Tc D T
(6.212)
The plots of sn , n D 1; : : : ; L, for L D 8 are shown in Figure 6.71. Choosing M D L, we note that also in this case the required bandwidth is proportional to 1=Tc D M=T , and we adopt the denition Bmin D M 2T (6.213)
Correspondingly, D .2 log2 M/=M, 0 D .2E s /=.N0 M/, and 0 I D 0. Example 6.7.4 (Binary code division modulation) We analyze in detail the previous example for M D 2 and L > M. In this case fsn .t/g n D 1; 2 (6.214)
coincide with two Walsh signals of length L and duration L Tc D T . Setting Bmin D L 2T (6.215)
from (6.93), (6.103), (6.105), and (6.106) we obtain, respectively, RI D D 0D 0I D 1 D 0:5 2 1 2 D .L=2T /T L 2E s N0 L Es N0 (6.216) (6.217) (6.218) (6.219)
Example 6.7.5 (Binary orthogonal modulation with coding) We have M D 2 as in the previous example. Now, however, the elements of the set fsn .t/g n D 1; 2 (6.220)
are chosen as two Walsh signals of length L and duration L Tc D 2T . Redundancy is introduced because the behavior of sn .t/ in the interval .T; 2T / depends on the behavior in
489
the interval .0; T /. With reference to the interval .0; 2T /, the number of information bits is equal to log2 M D 1: consequently, as T is the modulation interval we have L b D 1=2; note that we also have I D 2. Then, with respect to the binary code division modulation presented above, bandwidth and rate are halved, Bmin D L 4T 0:5 D 0:25 RI D 2 0:25 2 2D .L=4T /T L 4E s N0 L (6.221) (6.222)
We note that this case can be regarded as an example of a repetition code where the same symbol is repeated twice.
Probability of error
The ML receiver is given by the general scheme of Figure 6.8, where I D M and i is proportional to si according to (6.200). As the various signals have equal energy, the decision variables are given by Z t0 r.t/sn .t/ dt n D 1; : : : ; M (6.226) Un D Re[hr; sn i] D Re
0
t0 0
w.t/sn .t/ dt
(6.227)
E s wn
where wn D Re[hw; n i] is the n-th noise component. Then fUn g, n D 1; : : : ; M, are Gaussian r.v.s with mean mUn D E[Un ] D E s n and cross-covariance N0 E s ` n 2 Hence, the r.v.s fUn g are statistically independent with variance E s N0 =2. E[.Un mUn /.U` mU` /] D (6.229)
m
(6.228)
490
The probability of correct decision, conditioned on sm , is equal to P[C j sm ] D P[Um > U1 ; : : : ; Um > Um 1 ; Um > UmC1 ; : : : ; Um > U M ] Z a Z C1 Z a D pUm .a/ pU1 .b1 / : : : pU M .b M / db1 : : : db M da
1 1 1
1 Dp 2.E s N0 =2/
C1
e
1
1 .a E s /2 2 E s .N0 =2/
"
1 p 2.E s N0 =2/
e
1
b2 1 2 E s .N0 =2/ db
#M
da (6.230)
a E s N0 =2
1 2
Dp
r
b E s N0 =2
(6.231)
1 [1 Q./] M 1 d (6.232) p e 2 1 We note that (6.232) is independent of sm : consequently P[C j sm ] is the same for each sm . Therefore for equally likely signals we get P[C j sm ] D P[C] D P[C j sm ] The error probability is given by Pe D 1 P[C] Z
C1 1 2 r 2E s N0 !2
2E s N0
!2
(6.233)
(6.234) 1 M 1 D1 [1 Q./] d p e 2 1 Let M be a power of 2: with each signal sm we associate a binary representation, also called character, with log2 M bits. Then a signal error occurs if a character different from the transmitted character is detected. This error event happens with probability Pe . For each bit of the transmitted character, among the possible .M 1/ wrong characters only M=2 yield a wrong bit. Therefore we have M=2 Pbit D Pe M 1 (6.235) 1 ' Pe 2 for M sufciently large. Curves of Pbit as a function of 0 D 2E s =.N0 M/ and E b =N0 are given, respectively, in Figure 6.44 and Figure 6.45.15 We note that, in contrast with QAM modulation, for a given Pbit 0 decreases as M increases. The drawback is an increase of the required bandwidth with increasing M.
15 The computation of the integral (6.234) was carried out using the Hermite polynomial series expansion, as
491
10
10
10
Pbit
10
4
10
10 10
5 =2Es/(N0M) (dB)
10
15
20
Figure 6.44. Bit error probability as a function of 0 for transmission with M orthogonal signals.
10
10
10
Pbit
10
4
10
10
5
E / N (dB)
b 0
10
15
20
Figure 6.45. Bit error probability as a function of Eb =N0 for transmission with M orthogonal signals.
492
Figure 6.46. Comparison between the exact error probability and the limit (6.236) for transmission with M orthogonal signals.
(6.236)
Figure 6.46 shows a comparison between the error probability obtained by exact computation and the bound (6.236) for two values of M.
!0
M!1
1:59 dB
(6.238)
necessary to reach an error probability that can be made arbitrarily small for M ! 1 (see Section 6.10).
493
Table 6.8 Values of Eb =N0 required to obtain Pbit D 10 6 for various values of M.
E b =N0 (dB) 9.4 8.3 7.5 7.0 5.4 4.5 3.9 : : : 1:59
6.7.2
The elements of a set of M biorthogonal signals are M=2 orthogonal signals and their antipodal signals: for example, 4-PSK is a biorthogonal signalling scheme. A further example of biorthogonal signalling with 2M signals is given by a signalling scheme using the M orthogonal signals in (6.212) and their antipodal signals. For biorthogonal signalling with M signals, the required bandwidth is proportional to M=2. We give the parameters of the system in the two cases of non-coherent and coherent demodulation. Passband signalling with non-coherent demodulation: Bmin D M 1 2 T log2 M M (6.239) (6.240) (6.241)
D2 0D and, as I D M=2,
2E s N0 M
(6.242)
(6.243)
494
D4 0D and
log2 M M
(6.244) (6.245)
4E s N0 M
0I D 0
(6.246)
Probability of error
The receiver consists of M=2 correlators, or matched lters, which provide the decision variables fUn g n D 1; : : : ; M 2 (6.247)
The optimum receiver selects the output with the largest absolute value, jUi j; subsequently it selects si or si depending on the sign of Ui . To compute the probability of correct decision, we proceed as in the previous case. Assuming that sm is taken as one of the signals of the basis, then P[C j sm ] D P[Um > 0; jUm j > jU1 j; : : : ; jUm j > jUm > jUmC1 j; : : : ; jUm j > jU M=2 j] Z D
0 C1 1 j; jUm j
1 p e 2
1 2
2E s N0
!2
Curves of Pbit as a function of 0 D 4E s =.N0 M/ and E b =N0 are plotted, respectively, in Figure 6.47 and in Figure 6.48, for various values of M. A bound to (6.249) for transmission with M biorthogonal signals is given by s s ! ! Es 2E s CQ (6.251) Pe .M 2/Q N0 N0 where the rst term arises from the comparison with .M 2/ orthogonal signals, and the second arises from the comparison with an antipodal signal. Figure 6.49 shows a comparison between the error probability obtained by exact computation and the bound (6.251) for two values of M.
495
10
10
10
Pbit
10
4
10
10 10
=4Es/(N0M) (dB)
10
15
Figure 6.47. Bit error probability as a function of 0 for transmission with M biorthogonal signals.
10
10
10
Pbit
10
4
10
10
6 E / N (dB)
b 0
10
12
14
Figure 6.48. Bit error probability as a function of Eb =N0 for transmission with M biorthogonal signals.
496
Figure 6.49. Comparison between the exact error probability and the limit (6.251) for transmission with M biorthogonal signals.
6.8
We consider a baseband signalling scheme where the transmitted signal is given by sn .t/ D p Ew
n0 1 X jD0 T 1 Q where cn; j f 1; C1g, and wT .t/ D p rect t T =2 is the normalized rectangular window of T duration T (see (1.456)) with unit energy. Then E w is the energy of the pulse sn evaluated 1 on a generic subperiod T . Moreover, we have Bmin D 2T . Interpreting the n 0 pulses
cn; j wT .t Q
jT/
n D 1; : : : ; M
0 < t < n 0 T D Ts
(6.252)
wT .t/; : : : ; wT .t Q Q
.n 0
1/T /
(6.253)
as elements of an orthonormal basis, we derive the structure of the optimum receiver. Uncoded sequences. Every sequence of n 0 binary coefcients cn D [cn;0 ; : : : ; cn;n 0 is allowed, hence M D I D n 0 , it follows 2n 0 .
1] T
cn; j f 1; 1g
(6.254)
497
Es D n0 Ew Es D Ew I EI D Ew Eb D RI EI D 0I D 0D 2E w EI 2E b D D N0 =2 N0 N0 2E w Es D D 0I 1 N0 Ts N0 2T
Moreover, the minimum distance between two elements of the set of signals (6.252) is p equal to dmin D 4E w . The error probability is determined by the ratio (6.57)
u
(6.261)
where in the last step equation (6.258) is used. Coded sequences. We consider a set of signals (6.252) corresponding to M D 2k0 binary sequences cn with n 0 components, assuming that only k0 components in (6.254), as for example those with index j D 0; 1; : : : ; k0 1, can assume values in f 1; 1g arbitrarily: these components determine, through appropriate binary functions, also the remaining n 0 k0 components. Because the number of elements of the basis is always I D n 0 , we have RI D k0 log2 M D I n0 (6.262) (6.263) (6.264) (6.265) (6.266) (6.267)
Es D n0 Ew n0 Ew D Ew I EI n0 Eb D Ew D RI k0 EI D 0I D 0D 2E w EI k0 2E b D D N0 =2 N0 n 0 N0 Es D 0I 1 Ts N0 2T
H Indicating with dmin the minimum number of positions in which two vectors cn differ, we nd 2 H dmin D 4E w dmin
(6.268)
498
H In the case of uncoded sequences dmin D 1. An example of coding is given by the choice of the following vectors (code sequences or code words) for n 0 D 4 and k0 D 2, 2 3 2 3 2 3 2 3 1 1 C1 C1 6 17 6 17 6 C1 7 6 C1 7 7 7 7 7 c1 D 6 c2 D 6 c3 D 6 (6.269) c0 D 6 4 15 4 C1 5 4 15 4 C1 5 1 C1 1 C1 H 2 For this signalling system, we have dmin D 2, and therefore dmin D 8E w . Using (6.265), the signal-to-noise ratio at the decision point is given by
(6.270)
We note that for a given value of E b =N0 the coded system presents a larger , and H consequently a lower bit error probability, if dmin R I > 1. H We will discuss in Chapter 11 the design of codes that yield a large dmin for given values of the parameters n 0 and k0 . A drawback of these systems is represented by the reduction of the transmission bit rate Rb for a given modulation interval Ts ; alternative coding methods will be examined in Chapter 12.
Optimum receiver
With reference to the implementation of Figure 6.8, as the elements of the orthonormal basis (6.253) are obtained by shifting the pulse wT .t/, the optimum receiver can be simplied Q as illustrated in Figure 6.50, where the projections of the received signal r.t/ onto the
499
Figure 6.51. ML receiver for the signal set (6.252) under the assumption of uncoded sequences.
components of the basis (6.253) are obtained sequentially. The vector components r D [r0 ; r1 ; : : : ; rn o 1 ]T are then used to compute the Euclidean distances with each of the possible code sequences. The scheme of Figure 6.50 yields the detected signal of the type O O (6.252), or equivalently the detected code sequence c D [c0 ; c1 ; : : : ; cn o 1 ]T , according to O O the ML criterion. This procedure is usually called soft-input decoding. For the uncoded system, the receiver can be simplied by computing the Euclidean distance component by component, as illustrated in Figure 6.51. In the binary case under examination, Q ci D ci D sgn.ri / O i D 0; : : : ; n 0 1 (6.271)
The resulting channel model (memoryless binary symmetric) is that of Figure 6.21. In some receivers for coded systems, a simplication of the scheme of Figure 6.50 is obtained by rst detecting the single components ci f 1; 1g according to the scheme Q Q Q of Figure 6.51. Successively, the binary vector c D [c0 ; : : : ; cn 0 1 ]T is formed. Then we Q choose among the possible code sequences cn , n D 1; : : : ; 2k0 , the one that differs in the Q smallest number of positions with respect to the sequence c. This scheme is usually called hard -input decoding and is clearly suboptimum as compared to the scheme with soft input.
6.9
Table 6.9 summarizes some important results derived in the previous sections. Passband PAM is considered as single sideband (SSB) modulation or double sideband (DSB) modulation (see Appendix 7.C). In the latter case Bmin is equal to 1=T , hence 0 D E s =N0 . We note that, for a given noise level, PAM, PAMCSSB and PAMCDSB methods require the same statistical power to achieve a certain Pe ; however the PAMCDSB technique has a Bmin that is double as compared to PAM or PAMCSSB methods. For a given value of the symbol error probability, we now derive 0 I as a function of R I for some multilevel modulations. The result will be compared with the Shannon limit given by 0 I D 22R I 1, that represents the minimum theoretical value of 0 I , in correspondence of a given R I , for which Pbit can be made arbitrarily small by using channel coding without constraints in complexity and latency (see Section 6.10). We note
500
Table 6.9 Comparison of various modulation methods in terms of performance, bandwidth, and spectral efciency.
Modulation Approximated symbol error probability Pe p 0 r 2Q 2Q 4Q p 20 0 M r 1/Q r 2/Q M 0 2 1 T r 3 M2 6 1 3 1 M2 r M 1 0 0 ! 0 Minimum bandwidth Bmin (Hz) 1 2T 1 2T 1 T 1 T Spectral efciency D .1=Tb /=Bmin (bit/s/Hz) 2 Encoder- Signalmodulator to-noise rate R I ratio 0 (bit/dim) 1 2E s N0 2E s N0 Es N0 Es N0
binary antipodal (BB) M-PAM M-PAM C SSB M-PAM C DSB M-QAM .M D L 2 / BPSK o 2-PSK QPSK o 4-PSK M-PSK .M > 2/ 1 1 1 1 M 1 M
1 p M
log2 M 1 2 log2 M
r 2 sin2 2Q
Es N0
orthogonal (BB)
.M
M 2T M 4T
log2 M M log2 M M
2E s N0 M 4E s N0 M
biorthogonal (BB) .M
! ! r M M 0 CQ 0 4 2
that an equivalent approach often adopted in the literature is to give E b =N0 , related to 0 I through (6.107), as a function of , related to R I through (6.104). A rst comparison is made by assuming the same symbol error probability, Pe D 10 6 , p for all systems. As Q. z 0 / D 10 6 implies z 0 ' 22, considering only the argument of the Q function in Table 6.9, we have the following results. 1. M-PAM. From 3 M2 and 0I D 0 we obtain the following relation 0I D z 0 2R I .2 3 1/ (6.274) R I D log2 M (6.273) 1 0 D z0 (6.272)
501
0 D z0
(6.275)
log2 M
(6.276)
We note that for QAM a certain R I is obtained with a number of symbols equal to MQAM D 22R I , whereas for PAM the same efciency is reached for MPAM D p 2 R I D MQAM . 3. M-PSK. It turns out 0I D z 0 4R I 2 20 (6.278)
Equation (6.278) holds for M 4, and is obtained by approximating sin.=M/ with =M, and 2 with 10. 4. Orthogonal modulation. Using the approximation r Pe ' .M 1/Q ! (6.279)
M 0 2
we note that the multiplicative constant in front of the Q function cannot be ignored: therefore a closed-form analytical expression for 0 I as a function of R I for a given Pe cannot be found. 5. Biorthogonal modulation. The symbol error probability is approximately the same as that of orthogonal modulation for half the number of signals. Both R I and are doubled. We note that, for a given value of R I , PAM and QAM require the same value of 0 I , whereas PSK requires a much larger value of 0 I . An exact comparison is now made for a given bit error probability. Using the Pbit curves previously obtained, the behavior of R I as a function of 0 I for Pbit D 10 6 is illustrated in Figure 6.52. We observe that the required 0 I is much larger than the minimum value obtained by the Shannon limit. As will be discussed in Section 6.10, the gap can be reduced by channel coding. We also note that, for large R I , PAM and QAM allow a lower 0 I with respect to PSK; moreover, orthogonal and biorthogonal modulation operate with R I < 1, and corresponding very small values of 0 I .
502
Shannon limit
Figure 6.52. 0I required for a given rate RI , for different modulation methods and bit error probability equal to Pbit D 10 6 . The parameter in the gure denotes the number of symbols M of the constellation.
503
and depends mainly on the energy E s of the signal and on the spectral density N0 =2 of the noise. In addition to the required power and bandwidth, the choice of a modulation scheme is based on the channel characteristics and on the cost of the implementation: until recently, for example, non-coherent receivers were preferred in radio mobile systems because of their simplicity, even though the performance is inferior to that of coherent receivers (see Chapter 18) [2].
6.10
We consider the transmission of signals with a given power over an AWGN channel having noise power spectral density equal to N0 =2. We recall the denition (6.93) of the encoder-modulator rate, R I D L b =I , where I is the number of signal space dimensions. For example, the encoder-modulator for the 8-PAM system with bit map dened in Table 6.1 has rate R I D 3 (bit/dim), as L b D 3 and I D 1. From (6.95), we have the cardinality of alphabet A is equal to M D 2 R I D 8. Let us consider for example a monodimensional transmission system (PAM) with an alphabet of cardinality A for a given rate R I , such that L b < log2 M, that is M > 2 R I from (6.93); the redundancy of the alphabet can be used to encode sequences of information bits: in this case we speak of coded systems (see Example 6.7.5). Let us take a PAM system with R I D 3 and M D 16: redundancy may be introduced in the sequence of transmitted symbols. The mapping of sequences of information bits into sequences of coded output symbols may be described by a nite state sequential machine. Some specic examples will be illustrated in Chapter 12. We recall the denition (1.135) of the passband B associated with the frequency response R of a channel, with bandwidth given by (1.140), B D B d f . Channel capacity is dened as the maximum of the average mutual information between the input and output signals of the channel [6, 7]. For transmission over an ideal AWGN channel, channel capacity is given in bits per second by C[b=s] D B log2 .1 C 0/ (bit/s) (6.280)
where 0 is obtained from (6.105) by choosing Bmin D B. Equation (6.280) is a limit derived by Shannon assuming the transmitted signal s.t/ is a Gaussian random process with zero mean and constant power spectral density in the passband B. Using (6.280) and (6.103), we dene the maximum spectral efciency as max D C[b=s] D log2 .1 C 0/ (bit/s/Hz) B (6.281)
With reference to a message composed of a sequence of symbols, which belong to an I -dimensional space, the capacity can be expressed in bits per dimension as CD
1 2
log2 .1 C 0 I / (bit/dim)
(6.282)
504
obtained assuming a Gaussian distribution of the transmitted symbol sequence, where 0 I is given by (6.106). We give without proof the following fundamental theorem [8, 6]. Theorem 6.2 (Shannons theorem) For any rate R I < C, there exists channel coding that allows transmission of information with an arbitrarily small probability of error; such coding does not exist if R I > C. We note that Shannons theorem indicates the limits, in terms of encoder-modulator rate or, equivalently, in terms of transmission bit rate (see (6.280)), within which we can develop systems that allow reliable transmission of information, but it does not give any indication about the practical realization of channel coding. The capacity can be upper limited and approximated for small values of 0 I by a linear function, and also lower limited and approximated for large values of 0 I by a logarithmic function as follows: 0I 1 : C 0I 1 : C
1 2 1 2
(6.283) (6.284)
Extension of the capacity formula for an AWGN channel to multi-input multi-output (MIMO) systems can be found in [9, 10].
where pn indicates the probability of transmission of the symbol a0 D n . By the hypothesis of white Gaussian noise, we have ( ) . n /2 pr ja0 . j n / / exp (6.286) 2 I2 With the further hypothesis that only codes with equally likely symbols are of practical interest, the computation of the maximum of C with respect to the probability distribution of the input signal can be omitted. The channel capacity is therefore given by " !# 2 M Z M X 1 1 X C1 .n C i /2 2 2 I2 log Q D log2 M C e exp d p 2 M nD1 1 2 I2 2 I i D1 (6.287) Q is illustrated in Figure 6.53, where the Shannon limit given by (6.282), The capacity C as well as the signal-to-noise ratio given by (6.124) for which a symbol error probability
505
Figure 6.53. Capacity of an ideal AWGN channel for Gaussian and M-PAM input signals. [From Forney and Ungerboeck (1998). c 1998 IEEE.]
equal to 10 6 is obtained for uncoded transmission, are also indicated [12]. We note that the curves saturate as information cannot be transmitted with a rate larger than R I D log2 M. Let us consider, for example, the uncoded transmission of 1 bit of information per modulation interval by a 2-PAM system, where we have a symbol error probability equal to 10 6 for 0 I D 13:5 dB. If the number of symbols in the alphabet A doubles, choosing 4-PAM modulation, we see that the coded transmission of 1 bit of information per modulation interval with rate R I D 1 is possible, and an arbitrarily small error probability can be obtained for 0 I D 5 dB. This indicates that a coded 4-PAM system may achieve a gain of about 8:5 dB in signal-to-noise ratio over an uncoded 2-PAM system, at an error probability of 10 6 . If the number of symbols is further increased, the additional achievable gain is negligible. Therefore we conclude that, by doubling the number of symbols with respect to an uncoded system, we obtain in practice the entire gain that would be expected from the expansion of the input alphabet. We see from Figure 6.53 that for small values of 0 I the choice of a binary alphabet is almost optimum: in fact for 0 I < 1 (0 dB) the capacity given by (6.282) is essentially equivalent to the capacity given by (6.287) with a binary alphabet of input symbols. For large values of 0 I , the capacity of multilevel systems asymptotically approximates a straight line that is parallel to the capacity of the AWGN channel. The asymptotic loss of e=6 (1.53 dB) is due to the choice of a uniform distribution rather than Gaussian for the set of input symbols. To achieve the Shannon limit it is not sufcient to use coding techniques with equally likely input symbols, no matter how sophisticated they are: to bridge the gap
506
of 1.53 dB, shaping techniques are required [13] that produce a distribution of the input symbols similar to a Gaussian distribution. Coding techniques for small 0 I and large 0 I are therefore quite different: for low 0 I , the binary codes are almost optimum and the shaping of the constellation is not necessary; for high 0 I instead constellations with more than two elements must be used. To reach capacity, coding must be extended with shaping techniques; moreover, to reach the capacity in channels with limited bandwidth, techniques are required that combine coding, shaping and equalization, as we will see in Chapter 13.
(6.288)
for a given R I given by (6.93). For a scheme that achieves the capacity, R I is equal to the capacity of the channel C and 0 I D 1 (0 dB); if R I < C, as it must be in practice, then 0 I > 1. Therefore the value of 0 I indicates how far from the Shannon limit a system operates, or, in other words, the gap that separates the system from capacity. We now consider two cases. High signal-to-noise ratios. We note from Figure 6.53 that for high values of 0 I it is possible to nd coding methods that allow reliable transmission of several bits per dimension. For an uncoded M-PAM system, R I D log2 M (6.289)
bits of information are mapped into each transmitted symbol. The average symbol error probability is given by (6.124), ! r 3 1 Pe D 2 1 Q 0I (6.290) M M2 1 We note that Pe is function only of M and 0 I . Moreover, using (6.289) and (6.288) we obtain 0I D 0I M2 1 (6.291)
(6.292)
We note that the relation between Pe and 0 I is almost independent of M, if M is large. This relation is used in the comparison illustrated in Figure 6.54 between uncoded systems and the Shannon limit given by 0 I D 1.
507
Figure 6.54. Bit error probability as a function of Eb =N0 for an uncoded 2-PAM system, and symbol error probability as a function of 0 I for an uncoded M-PAM system. [From Forney and Ungerboeck (1998). c 1998 IEEE.]
Low signal-to-noise ratios. For low values of 0 I the capacity is less than 1 and can be approximated by binary transmission systems: consequently we refer to coding methods that employ more binary symbols to obtain the reliable transmission of 1 bit (see Section 6.8). For low values of 0 I it is customary to introduce the following ratio (see (6.107)): 22R I 1 Eb D 0I N0 2R I We note the following particular cases: if R I 1, then E b =N0 .ln 2/ 0 I ; if R I D 1=2, then E b =N0 D 0 I ; if R I D 1, then E b =N0 D .3=2/ 0 I . For low 0 I , if the bandwidth can be extended without limit for a given power, for example, by using an orthogonal modulation with T ! 0 (see Example 6.7.3), then by increasing the bandwidth, or equivalently the number of dimensions M of input signals, both 0 I and R I tend to zero. For systems with limited power and unlimited bandwidth, usually E b =N0 is adopted as a gure of merit. (6.293)
508
From (6.293) and the Shannon limit 0 I > 1, we obtain the Shannon limit in terms of E b =N0 for a given rate R I as Eb 22R I 1 > 2R I N0 This lower limit monotonically decreases with R I . In particular, we examine again the three cases: if R I tends to zero, the ultimate Shannon limit is given by Eb > ln 2 N0 . 1:59 dB/ (6.295) (6.294)
in other words, equation (6.295) afrms that even though an innitely large bandwidth is used, reliable transmission can be achieved only if E b =N0 > 1:59 dB; if the bandwidth is limited, from (6.294) we nd that the Shannon limit in terms of E b =N0 is higher; for example, if R I D 1=2 the limit becomes E b =N0 > 1 (0 dB); if R I D 1, as E b =N0 D .3=2/ 0 I , the symbol error probability or bit error probability for an uncoded 2-PAM system can be expressed in two equivalent ways: s ! q 2E b Pbit Q 30 I D Q (6.296) N0
Coding gain
Denition 6.2 The coding gain of a coded modulation scheme is equal to the reduction in the value of E b =N0 , or in the value of 0 or 0 I (see (11.9)), that is required to obtain a given probability of error relative to a reference uncoded system. If the modulation rate of the coded system remains unchanged, we typically refer to 0 or 0 I . Let us consider as reference systems a 2-PAM system and an M-PAM system with M 1, for small and large values of 0 I , respectively. Figure 6.54 illustrates the bit error probability for an uncoded 2-PAM system as a function of both E b =N0 and 0 I . For Pbit D 10 6 , the reference uncoded 2-PAM system operates at about 12.5 dB from the ultimate Shannon limit. Thus a coding gain up to 12.5 dB is possible, in principle, at this probability of error, if the bandwidth can be sufciently extended to allow the use of binary codes with R I 1; if, instead, the bandwidth can be extended only by a factor 2 with respect to an uncoded system, then a binary code with rate R I D 1=2 can yield a coding gain up to about 10.8 dB. Figure 6.54 also shows the symbol error probability for an uncoded M-PAM system as a function of 0 I for large M. For Pe D 10 6 , a reference uncoded M-PAM system operates at about 9 dB from the Shannon limit: in other words, assuming a limited bandwidth system, the Shannon limit can be achieved by a code having a gain of about 9 dB.
509
Cut-off rate
It is useful to introduce the notion of cut-off rate R0 associated with a channel, for a given modulation and class of codes [2]. We sometimes refer to R0 as a practical upper bound of the transmission bit rate. Therefore for a given channel we can determine the minimum signal-to-noise ratio .E b =N0 /0 below which reliable transmission is not possible, assuming a certain class of coding and decoding techniques. Typically, for codes with rate Rc D 1 (see Chapter 11), .E b =N0 /0 is 2 about 2 dB above the signal-to-noise ratio at which capacity is achieved.
6.11
Let us consider transmission over an AWGN channel of one of the signals n D 1; 2; : : : ; M (6.297)
.bb/ where sn is the complex envelope of sn , relative to the carrier frequency f 0 , with support .bb/ .0; t0 /. If in (6.297) every signal sn has a bandwidth smaller than f 0 , then the energy of sn is given by Z t0 Z t0 1 .bb/ 2 2 jsn .t/j dt En D sn .t/ dt D (6.298) 0 0 2
(6.299)
(6.300) ] n D 1; 2; : : : ; M
j2 f 0 t
In other words, at the receiver we assume the carrier is known, except, however, for a phase ' that we assume to be a uniform r.v. in [ ; /. Receivers, which do not rely on the knowledge of the carrier phase, are called non-coherent receivers. We give three examples of signalling schemes that employ non-coherent receivers. Example 6.11.1 (Non-coherent binary FSK) The received signals are expressed as (see also (6.204)): s1 .t;'1 / D A cos.2 f 1 t C '1 / s2 .t;'2 / D A cos.2 f 2 t C '2 / where r AD 2E s T '1 ; '2 2 U [ ; / (6.302) 0<t <T 0<t <T (6.301)
510
and f1 D f0 fd f2 D f0 C fd (6.303)
where f d is the frequency deviation with respect to the carrier f 0 . We recall that if f 1 C f 2 D k1 and if 2 f d T D k (k integer) then s1 .t; '1 / and s2 .t; '2 / are orthogonal. The minimum value of f d is given by . f d /min D 1 2T (6.306) (6.305) 1 (k1 integer) T or else f0 1 T (6.304)
which is twice the value we nd for the coherent demodulation case (6.203). Example 6.11.2 (On-off keying) On-off keying (OOK) is a binary modulation scheme where, for example, s1 .t;'/ D A cos.2 f 0 t C '/ and s2 .t;'/ D 0 where A D p 4E s =T , and E s is the average energy of a pulse. (6.308) 0<t <T (6.307)
Example 6.11.3 (DSB modulated signalling with random phase) .bb/ We consider an M-ary baseband signalling scheme, fsn .t/g, n D 1; : : : ; M, that is modulated in the passband by the double sideband technique (see Example 1.7.3 on page 58). The received signals are expressed as
.bb/ sn .t;'/ D sn .t/ cos.2 f 0 t C '/
n D 1; : : : ; M
(6.309)
ML criterion
Given ' D p, that is for known ', the ML criterion to detect the transmitted signal has been previously developed starting from (6.26). The conditional probability density function of the vector r is given by
1 1 jj sn jj2 prja0 ;' . j n; p/ D p e N0 I . 2.N0 =2// Z t0 2 D K exp r.t/ sn .t; p/ dt N0 0
1 N0
Z
0
t0
2 sn .t; p/ dt
(6.310)
511
(6.311)
we dene the following likelihood function, which is equivalent, but not equal, to that dened in (6.27): Z t0 2 En exp Ln [ p] D exp r.t/ sn .t; p/ dt n D 1; : : : ; M (6.312) N0 N0 0 Given ' D p, the maximum likelihood criterion yields the decision rule a0 D arg max Ln [ p] O
n
(6.313)
The dependency on the r.v. ' is removed by taking the expectation of Ln [ p] with respect to ':16 Z Ln D Ln [ p] p' . p/ d p
De
En N0
1 2
exp
2 Re N0
Z
0
t0
dp
(6.314)
(6.315)
1 2 1 2
Z Z
p 2 En Re[L n e e N0
jp]
dp (6.316)
De
En N0
p 2 En jL j cos. p arg L n / e N0 n dp
e x cos. p
dp
(6.317)
16 Averaging with respect to the phase ' cannot be considered for PSK and QAM systems, where information is
512
p 2 En I0 jL n j N0
n D 1; : : : ; M
(6.318)
(6.319)
If the signals have all the same energy, and considering that both ln and I0 are monotonic functions, the ML decision criterion can be expressed as a0 D arg max jL n j O
n
(6.320)
(6.321)
From (6.315), the scheme rst determines the real and the imaginary parts of L n starting .bb/ from sn , and then determines the squared magnitude. Note that the available signal is .bb/ .bb/ sn .t/ e j'0 , where '0 is a constant, rather than sn .t/: this, however, does not modify the magnitude of L n . As shown in Figure 6.56, the generic branch of the scheme in Figure 6.55, composed of the I branch and the Q branch, can be implemented by a complex-valued passband lter (see (6.315)); the bold line denotes a complex-valued signal. Alternatively, the matched lter can be real valued if it is followed by a phase-splitter: in this case the receiver is illustrated in Figure 6.57. For the generic branch, the desired value jL n j coincides with the absolute value of the output signal of the phase-splitter at instant t0 ,
.a/ jL n j D jyn .t/jtDt0
(6.322)
The cascade of the phase-splitter and the modulo transformation is called the envelope detector of the signal yn .t/ (see (1.196) and (1.202)). .bb/ A simplication arises if the various signals sn have a bandwidth B much lower than .bb/ f 0 . In this case, recalling (1.202), if yn is the complex envelope of yn , at the matched lter output the following relation holds
.bb/ yn .t/ D Re[yn .t/e j2 f 0 t ] .bb/ .bb/ D jyn .t/j cos.2 f 0 t C arg yn .t//
(6.323)
(6.324)
513
r(t)
|L n | 2
Figure 6.56. Implementation of a branch of the scheme of Figure 6.55 by a complex-valued passband matched lter.
.bb/ Now, if f 0 B, to determine the amplitude jyn .t/j, we can use one of the schemes of Figure 6.58.
Example 6.11.4 (Non-coherent binary FSK) We show in Figure 6.59 two alternative schemes of the ML receiver for the modulation system considered in Example 6.11.1.
514
Figure 6.57. Non-coherent ML receiver of the type envelope detector, using passband matched lters.
(a)
(b)
Figure 6.58. (a) Ideal implementation of an envelope detector, and (b) two simpler approximate implementations.
Example 6.11.5 (On-off keying) We illustrate in Figure 6.60 the receiver for the modulation system of Example 6.11.2 where, recalling (6.318), we have N0 UTh D p I0 1 .e E 1 =N0 / 2 E1 where E1 D A2 T 2 (6.326) (6.325)
515
r(t)
T w (t)cos(2 f t+ )
T 0 0
envelope detector
^ U U>UTh , a0 =1 ^ U<UTh , a =2 0
^ a
516
Example 6.11.6 (DSB modulated signalling with random phase ) With reference to the Example 6.11.3, we show in Figure 6.61 the receiver for a baseband M-ary signalling scheme that is DSB modulated with random phase. Depending upon the signalling type, further simplications can be done by extracting functions that are common to the different branches.
(6.327)
(6.328)
p U1
and
V2 D
p U2
(6.329)
(6.330)
2 C
Z
0
2 (6.331)
D where
2 w2;c
2 C w2;s
Z
0
Z
0
Z
0
Z
0
517
Figure 6.61. Two receivers for a DSB modulation system with M-ary signalling and random phase.
518
we have Z U1 D
0 T
2 C C
Z
0
2 (6.334)
AT cos.'0 2
AT sin.'0 2
r AT Es T D (6.335) 2 2 As w.t/ is a white Gaussian random process with zero mean, w2;c and w2;s are two jointly Gaussian r.v.s with E[w2;c ] D E[w2;s ] D 0
2 2 E[w2;c ] D E[w2;s ] D
E[w2;c w2;s ] D
Z
0
Z
0
N0 .t1 2
(6.336) Similar considerations hold for w1;c and w1;s . Therefore V2 , with statistical power 2.N0 T =4/, has a Rayleigh probability density v2 e pV2 .v2 / D N0 T =4
2 v2 2.N0 T =4/ 1.v
2/
(6.337)
I0
(6.338)
Consequently equation (6.330) assumes the expression17 Z C1 Pbit D P[V1 < v2 j V2 D v2 ] pV2 .v2 / dv2
0
Z D
0
C1 Z v2 0 1 Es 2 N0
(6.340)
1 e 2
Z C1
0
2 1 2 x e .x =2/ I0 .x/ d x D e =2
(6.339)
519
It can be shown that this result is not limited to FSK systems and is valid for any pair of non-coherent orthogonal signals with energy E s .
FSK(NC):
Pbit D
DBPSK:
Pbit D 1 e 2
Es N0
(6.342)
A comparison between (6.341) and (6.342) indicates that DBPSK is better than FSK by about 3 dB in 0, for the same Pbit . The performance of a binary FSK system and that of a differentially encoded BPSK system with coherent detection are compared in Figure 6.62. In particular, from (6.75) it follows s ! Es FSK(CO): Pbit D Q (6.343) N0
10
1
10
10
FSK ( =0)
FSK NC
Pbit
CO
4
10
(d.e.)BPSK ( =1)
10
5
DBPSK
10
10 =E /N (dB)
s 0
11
12
13
14
15
Figure 6.62. Bit error probability as a function of 0 for BPSK and binary FSK systems, with coherent (CO) and non-coherent (NC) detection.
520
Taking into account differential decoding, from (6.64) we have18 s ! 2E s (d.e.)BPSK: Pbit ' 2Q N0
(6.344)
We observe that the difference between coherent FSK and non-coherent FSK is less than 2 dB for Pe 10 3 , and becomes less than 1 dB for Pe 10 5 . We also note that because of the large bandwidth required, see (6.206), FSK systems with M > 2 are not widely used.
6.12
We assume now that the channel introduces, in addition to a random phase, also a random attenuation (see Section 4.6.5). The received signal is expressed as
.bb/ r.t/ D Re[sn .t/g1 e j2 f 0 t ] C w.t/
n 2 f1; : : : ; Mg
0<t <T
(6.345)
where g1 D g1;I C jg1;Q is a Rayleigh r.v., that is g1;I and g1;Q are uncorrelated Gaussian r.v.s, with zero mean and equal variance. In polar notation g1 D jg1 je j' , where ' 2 U . ; / and pjg1 j is given by (4.215). As jg1 j determines the signal level at the input of the receiver, the signal-to-noise ratio at the receiver input, 0D Es jg1 j2 N0 (6.346)
where E s is the average energy of the transmitted signal, is a function of jg1 j and therefore it is a random variable. We dene the average signal-to-noise ratio 0avg D Es E[jg1 j2 ] N0 1 0avg (6.347)
1.a/
(6.348)
To compute the performance of a signalling scheme in the presence of at fading, we regard the expressions of Pe obtained in the previous sections as functions of 0. Therefore we consider Pe as the conditional error probability for a given value of jg1 j. To evaluate the mean error probability we apply the total probability theorem, that yields Z C1 Pe D Pe .a/ p0 .a/ da (6.349)
0
Limiting ourselves to binary signalling schemes and substituting Pe given by (6.341), (6.342), (6.343) and (6.344) in (6.349), we obtain:19
18 For a more accurate evaluation of the probability of error see footnote 14 on page 478. 19 For the computation of the integral in (6.349) we recall the following result:
Z C1 Q
0
x p 1 1 x dx D e 2
s 1
2 C
! (6.350)
521
1. Orthogonal binary FSK with coherent detection s ! 0avg 1 Pbit D 1 2 2 C 0avg 2. Differentially encoded BPSK with coherent detection s 0avg Pbit D 1 1 C 0avg
(6.351)
(6.352)
We note that both the above expressions are in practice a lower limit to Pbit , as it is assumed that an estimate of the phase ' is available, which is very hard to obtain under fading conditions. In case the uncertainty on the phase is relevant, non-coherent DPSK and FSK systems are valid alternatives. 3. Orthogonal binary FSK with non-coherent detection Pbit D 4. DBPSK Pbit D 1 2.1 C 0avg / (6.354) 1 2 C 0avg (6.353)
The various expressions of Pbit as a function of 0avg are plotted in Figure 6.63 and compared with the case of transmission over an AWGN channel. We note that to achieve a certain Pbit , it is required a substantially larger E s as compared to the case of transmission over an AWGN channel, for the same N0 . For a systematic method to determine the performance of systems in the presence of a channel affected by multipath fading, we refer the reader to [14], and to the references therein.
Diversity
In the previous section it became apparent that the probability of error for transmission over channels with Rayleigh fading varies inversely proportional to the signal-to-noise ratio, rather than exponentially as in the AWGN channel case: therefore a large transmission power is needed to obtain good system performance. To mitigate this problem it is useful to resort to the concept of diversity, that is exploiting channels that are independent, or at least highly uncorrelated, for communication. The basic idea consists in providing the receiver with several replicas of the signal via independent channels, so that the probability is small that the attenuation due to fading is high for all the channels. There are various diversity techniques. 1. Frequency diversity: the same signal is transmitted using several carriers, separated from each other in frequency by an interval that is larger than the coherence bandwidth of the channel.
522
Figure 6.63. Bit error probability as a function of 0avg for BPSK, DBPSK, and binary FSK systems, for a at Rayleigh fading channel.
2. Time diversity: the same signal is transmitted over different time slots, spaced by an interval that is larger than the coherence time of the channel. 3. Space diversity: multiple reections from ground and surrounding buildings can make the power of the received signal change rapidly; by setting two or more antennas close to each other, we can select the antenna that provides the signal with higher power. 4. Polarization diversity: several channels are obtained for transmission by orthogonal polarization. 5. Combinations of the previous techniques: for the many techniques of combining available, both linear (equal gain, selection, maximal ratio) and non-linear (square law ), we refer to Section 8.18 and to the bibliography [15, 16, 17].
6.13
6.13.1
Transmission methods
Transmission methods between two users
A transmission link between two users of a communication network may be classied as a) Full duplex, when two users A and B can send information to each other simultaneously, not necessarily by using the same transmission channels in the two directions.
523
b) Half duplex, when two users A and B can send information in only one direction at a time, from A to B or from B to A, alternatively. c) Simplex, when only A can send information to B, that is the link is unidirectional.
Three methods
In the following we give three examples of transmission methods which are used in practice. a) Frequency-division duplexing (FDD): in this case the two users are assigned different transmission bands using the same transmission medium, thus allowing full-duplex transmission. Examples of FDD systems are the GSM, which uses a radio channel (see Section 17.A.2), and the VDSL, which uses a twisted pair cable (see page 1146). b) Time-division duplexing (TDD): in this case the two users are assigned different slots in a time frame (see Section 6.13.2). If the duration of one slot is small with respect to that of the message, we speak of full-duplex TDD systems. Examples of TDD systems are the DECT, which uses a radio channel (see Section 17.A.6), and the ping-pong BR-ISDN, which uses a twisted pair cable. c) Full-duplex systems over a single band: in this case the two users transmit simultaneously in two directions using the same transmission band; examples are the HDSL (see Section 17.1.1), and in general high-speed transmission systems over twisted-pair cables for LAN applications (see Section 17.1.2). The two directions of transmission are separated by a hybrid; the receiver eliminates echo signals by echo cancellation techniques. We note that full-duplex transmission over a single band is possible also over radio channels, but in practice alternative methods are still preferred because of the complexity required by echo cancellation.
6.13.2
We distinguish three cases for channel access by N users: 1. Subdivision of the channel passband into N B separate sub-bands that may be used for transmission (see Figure 6.64a). 2. Subdivision of a sequence of modulation intervals into adjacent subsets called frames, each in turn subdivided into N S adjacent subsets called slots. Within a frame, each slot is identied by an index i, i D 0; : : : ; N S 1 (see Figure 6.64b). 3. Signalling by N0 orthogonal signals (see for example Figure 6.71). N users may share the channel using one of the following methods.20 1. Frequency division multiple access (FDMA): to each user is assigned one of the N B sub-bands.
20 The access methods discussed in this section are deterministic, as each user knows exactly at which point in
time the channel resources are reserved for transmission; an alternative approach is represented by random access techniques, e.g., ALOHA, CSMA/CD, collision resolution protocols [18] (see also Chapter 17).
524
2. Time division multiple access (TDMA): to each user is assigned one of the N S time sequences (slots), whose elements identify the modulation intervals. 3. Code division multiple access (CDMA): to each user is assigned a modulation scheme that employs one of the N0 orthogonal signals, preserving the orthogonality between modulated signals of the various users. For example, for the case N0 D 8, to each user may be assigned one orthogonal signal of those given in Figure 6.71; for binary modulation, within a modulation interval each user then transmits the assigned orthogonal signal or the antipodal signal. We give an example of implementation of the TDMA principle. Example 6.13.1 (Time-division multiplexing) Time-division multiplexing (TDM) is the interleaving of several digital messages into one digital message with a higher bit rate; as an example we illustrate the generation of the European base group, called E1, at 2.048 Mbit/s, that is obtained by multiplexing 30 PCM coded speech signals (or channels) at 64 kbit/s. As shown in Figure 6.65, each 8-bit sample of each channel is inserted into a pre-assigned slot of a frame composed of 32 slots, equivalent to 328 D 256 bits. The frame structure must contain information bits to identify the beginning of a frame (channel ch0) by 8 known framing bits; 8 bits are employed for signalling between central ofces (channel ch16). The remaining 30 channels are for the transmission of signals. As the duration of a frame is of 125 s, equal to the interval between two PCM samples of the same channel, the overall digital message has a bit rate of 256 bit/125 s = 2.048 Mbit/s; we note, however, that of the 256 bits of the frame only 30 8 D 240 bits carry information related to signals. In the United States, Canada, and Japan the base group, analog to E1, is called T1 carrier system and has a bit rate of 1.544 Mbit/s, obtained by multiplexing 24 PCM speech coded signals at 64 kbit/s. In this case the frame is such that one bit per channel is employed for signalling: this bit is robbed from the least important bit of the 8-bit PCM sample, thus making it a 7-bit code word per sample; there is then only one bit for the synchronization of the whole frame. The entire frame is formed of 24 8 C 1 D 193 bits.
6. Bibliography
525
Bibliography
[1] J. M. Wozencraft and I. M. Jacobs, Principles of communication engineering. New York: John Wiley & Sons, 1965. [2] S. Benedetto and E. Biglieri, Principles of digital transmission with wireless applications. New York: Kluwer Academic Publishers, 1999. [3] J. G. Proakis, Digital communications. New York: McGraw-Hill, 3rd ed., 1995. [4] D. Divsalar, M. K. Simon, and M. Shahshahani, The performance of trellis-coded MDPSK with multiple symbol detection, IEEE Trans. on Communications, vol. 38, pp. 13911403, Sept. 1990. [5] M. Abramovitz and I. A. Stegun, eds, Handbook of mathematical functions. New York: Dover Publications, 1965.
526
[6] R. G. Gallager, Information theory and reliable communication. New York: John Wiley & Sons, 1968. [7] T. M. Cover and J. Thomas, Elements of information theory. New York: John Wiley & Sons, 1991. [8] C. E. Shannon, A mathematical theory of communication, Bell System Technical Journal, vol. 27, pp. 379427 (Part I) and 623656 (Part II), 1948. [9] G. J. Foschini and M. J. Gans, On limits of wireless communications in a fading environment when using multiple antennas, Wireless Person. Commun., vol. 6, pp. 311335, June 1998. [10] E. Telatar, Capacity of multiantenna Gaussian channels, Europ. Trans. on Telecomm., vol. 10, pp. 585595, Nov.Dec. 1999. [11] G. Ungerboeck, Channel coding with multilevel/phase signals, IEEE Trans. on Information Theory, vol. 28, pp. 5567, Jan. 1982. [12] G. D. Forney, Jr. and G. Ungerboeck, Modulation and coding for linear Gaussian channels, IEEE Trans. on Information Theory, vol. 44, pp. 23842415, Oct. 1998. [13] G. D. Forney, Jr., Trellis shaping, IEEE Trans. on Information Theory, vol. 38, pp. 281300, Mar. 1992. [14] M. K. Simon and M.-S. Alouini, Exponential-type bounds on the generalized Marcum Q-function with application to error probability analysis over fading channels, IEEE Trans. on Communications, vol. 48, pp. 359366, Mar. 2000. [15] M. Schwartz, W. R. Bennett, and S. Stein, Communication systems and techniques. New York: McGraw-Hill, 1966. [16] T. S. Rappaport, Wireless communications: principles and practice. Englewood Cliffs, NJ: Prentice-Hall, 1996. [17] G. L. Stuber, Principles of mobile communication. Norwell, MA: Kluwer Academic Publishers, 1996. [18] Multiple access communications: foundations for emerging technologies. (N. Abramson, ed.), Piscataway: IEEE Press, 1993. [19] W. F. McGee, Another recursive method of computing the Q-function, IEEE Trans. on Information Theory, vol. 16, pp. 500501, July 1970. [20] R. E. Ziemer and W. H. Tranter, Principles of communications: systems, modulation, and noise. New York: John Wiley & Sons, 4th ed., 1995.
527
Appendix 6.A
6.A.1
The Q function
The probability density function of a Gaussian variable w with mean m and variance 2 is given by e pw .b/ D p 2 1
.b m/2 2 2
(6.355)
We dene normalized Gaussian distribution (m D 0 and 2 D 1) as the function Z a Z a 1 2 8.a/ D pw .b/ db D p e b =2 db (6.356) 2 1 1 It is often convenient to use the complementary Gaussian distribution function, dened as Z C1 1 2 p e b =2 db (6.357) Q.a/ D 1 8.a/ D 2 a Two other functions that are widely used are the error function erf .a/ D D1 and the complementary error function erfc .a/ D 1 erf .a/ (6.359) Z 1C2
1 p a 2
pw .b/ db (6.358)
b2
Z 2
C1
p a 2
1 p e 2
db
which are related to 8 and Q by the following equations a 1 8.a/D 1 C erf p 2 2 1 a Q.a/D erfc p 2 2
(6.360) (6.361)
528
a 0:0 0:1 0:2 0:3 0:4 0:5 0:6 0:7 0:8 0:9 1:0 1:1 1:2 1:3 1:4 1:5 1:6 1:7 1:8 1:9 2:0 2:1 2:2 2:3 2:4 2:5 2:6
Q.a/ 5:0000. 4:6017. 4:2074. 3:8209. 3:4458. 3:0854. 2:7425. 2:4196. 2:1186. 1:8406. 1:5866. 1:3567. 1:1507. 9:6800. 8:0757. 6:6807. 5:4799. 4:4565. 3:5930. 2:8717. 2:2750. 1:7864. 1:3903. 1:0724. 8:1975. 6:2097. 4:6612. 01/ 01/ 01/ 01/ 01/ 01/ 01/ 01/ 01/ 01/ 01/ 01/ 01/ 02/ 02/ 02/ 02/ 02/ 02/ 02/ 02/ 02/ 02/ 02/ 03/ 03/ 03/
a 2:7 2:8 2:9 3:0 3:1 3:2 3:3 3:4 3:5 3:6 3:7 3:8 3:9 4:0 4:1 4:2 4:3 4:4 4:5 4:6 4:7 4:8 4:9 5:0 5:1 5:2 5:3
Q.a/ 3:4670. 2:5551. 1:8658. 1:3499. 9:6760. 6:8714. 4:8342. 3:3693. 2:3263. 1:5911. 1:0780. 7:2348. 4:8096. 3:1671. 2:0658. 1:3346. 8:5399. 5:4125. 3:3977. 2:1125. 1:3008. 7:9333. 4:7918. 2:8665. 1:6983. 9:9644. 5:7901. 03/ 03/ 03/ 03/ 04/ 04/ 04/ 04/ 04/ 04/ 04/ 05/ 05/ 05/ 05/ 05/ 06/ 06/ 06/ 06/ 06/ 07/ 07/ 07/ 07/ 08/ 08/
a 5:4 5:5 5:6 5:7 5:8 5:9 6:0 6:1 6:2 6:3 6:4 6:5 6:6 6:7 6:8 6:9 7:0 7:1 7:2 7:3 7:4 7:5 7:6 7:7 7:8 7:9 8:0
Q.a/ 3:3320. 1:8990. 1:0718. 5:9904. 3:3157. 1:8175. 9:8659. 5:3034. 2:8232. 1:4882. 7:7688. 4:0160. 2:0558. 1:0421. 5:2310. 2:6001. 1:2798. 6:2378. 3:0106. 1:4388. 6:8092. 3:1909. 1:4807. 6:8033. 3:0954. 1:3945. 6:2210. 08/ 08/ 08/ 09/ 09/ 09/ 10/ 10/ 10/ 10/ 11/ 11/ 11/ 11/ 12/ 12/ 12/ 13/ 13/ 13/ 14/ 14/ 14/ 15/ 15/ 15/ 16/
In Table 6.10 the values assumed by the complementary Gaussian distribution are given for values of the argument between 0 and 8. We present below some bounds of the Q function. 2 1 a 1 1 exp bound1 : Q 1 .a/ D p (6.362) 2 2 a 2a 2 a 1 bound2 : Q 2 .a/ D p exp (6.363) 2 2a 2 a 1 (6.364) bound3 : Q 3 .a/ D exp 2 2 The Q function and the above bounds are illustrated in Figure 6.66.
529
6.A.2
(6.365)
where I0 is the modied Bessel function of the rst type and order zero, dened in (4.216). From (6.365), two particular cases follow: Q 1 .0; b/De Q 1 .a; 0/D1 Moreover, for b 1 and b b a the following approximation holds Q 1 .a; b/ ' Q.b a/ a > 0, is given by a/ (6.369) (6.368)
b2 2
(6.366) (6.367)
where the Q function is given by (6.357). A useful approximation valid for b 1, a 1, b b 1 C Q 1 .a; b/ We also give the Simon bound [14] e
.bCa/2 2
Q 1 .a; b/ e
.b a/2 2
b>a>0
(6.370)
530
and 1 " 1 e 2
.a b/2 2
.aCb/2 2
We observe that in (6.370) the upper bound is very tight, and the lower bound for a given value of b becomes looser as a increases. In (6.371) the lower bound is very tight. A recursive method for computing the Marcum function is given in [19].
531
Appendix 6.B
Gray coding
In this appendix we give the procedure to construct a list of 2n binary words of n bits, where adjacent words differ in only one bit. The case for n D 1 is immediate. We have two words with two possible values 0 1 (6.372)
The list for n D 2 is constructed by considering rst the list of .1=2/22 D 2 words that are obtained by appending a 0 in front of the words of the list (6.372): 0 0 0 1 (6.373)
The remaining two words are obtained by inverting the order of the words in (6.372) and appending a 1 in front: 1 1 1 0 The nal result is the following list of words: 0 0 1 1 0 1 1 0 (6.374)
(6.375)
Iterating the procedure for n D 3, the rst 4 words are obtained by repeating the list (6.375) and appending a 0 in front of the words of the list. Inverting then the order of the list (6.375) and appending a 1 in front, the nal result is the list of 8 words 0 0 0 0 1 1 1 1 0 0 1 1 1 1 0 0 0 1 1 0 0 1 1 0
(6.376)
It is easy to extend this procedure to any value of n. By induction it is just as easy to prove that two adjacent words in each list differ by one bit at most.
532
Appendix 6.C
In addition to the widely known PAM, two other baseband pulse modulation techniques are pulse position modulation (PPM) and pulse duration modulation (PDM). PPM consists of a set of pulses whose shift, with respect to a given time reference, depends on the value of the transmitted symbol. We consider the fundamental pulse shape of Figure 6.67 and an alphabet given by A D f0; 1; 2; 3; : : : ; M The transmitted isolated pulse is sn .t/ D g0 t T n M n2A (6.378) 1g (6.377)
For M D 4 the set of waveforms is represented in Figure 6.68. In PDM, instead, the input symbol determines the duration of the transmitted pulse, that is a multiple of a minimum time duration equal to T =M. For an alphabet A D f1; 2; : : : ; Mg the transmitted isolated pulse is given by sn .t/ D g0 t n n2A (6.380) (6.379)
where g0 is given in Figure 6.67. The set of PDM waveforms for M D 4 is illustrated in Figure 6.69.
Signal-to-noise ratio
In both PPM and PDM the information lies in the position of the fronts of the transmitted pulses: therefore demodulation consists in nding the fronts of the pulses, which are disturbed by noise. If the channel bandwidth were innite, one could receive perfectly rectangular pulses. In practice the received pulse, with amplitude equal to A, has a rise time t R different from
g (t)
0
0 T/M
T t
533
n =1
T/4
t n =2
2T/4
t n =3
3T/4
t n =4
T t
n =1
T/4
t n=2
2T/4
t n=3
3T/4
t n =4
534
zero, as the channel has a nite bandwidth B, and noise disturbs the reception of the pulse, as illustrated in Figure 6.70. The detection of the front of a pulse is obtained by establishing the instant ti in which the received signal, pulse plus noise, crosses a given threshold. The error ".ti / is related to the noise w.ti /, amplitude A, and rise time t R of the received pulse: w.ti / ".ti / D tR A (6.381)
Assuming the noise stationary with PSD N0 =2 over the channel passband with bandwidth B, the mean-square error is given by E[" ] D
2
tR A
E[w ] D
tR A
2 N0 B (6.382)
We consider the following approximated expression that links the rise time to the bandwidth of the pulse tR ' 1 2B (6.383)
Substitution of the above result in (6.382) yields E[" 2 ] ' N0 4A2 B (6.384)
On the other hand, assuming the average duration of the pulses is 0 , the signal-to-noise ratio is given by (6.105) with Bmin D 1=.2T /, that is 0D 2E sCh N0 (6.385)
535
where E sCh D 0 A2 Finally, substitution of (6.385) in (6.384) yields E[" 2 ] D 0 1 2B 0 (6.387) (6.386)
For a more in-depth analysis we refer to [20], where a trade-off between the channel bandwidth and the signal-to-noise ratio at the decision point (see (6.387)) is illustrated.
536
Appendix 6.D
Walsh codes
We illustrate a procedure to obtain orthogonal binary sequences, with values f 1; 1g, of length 2m . We consider 2m 2m Hadamard matrices Am , with binary elements from the set f0; 1g. For the rst orders, we have A0D[0] 0 0 A1D 0 1 (6.388) (6.389)
Figure 6.71. Eight orthogonal signals obtained from the Walsh code of length 8.
537
0 1 0 1 0 1 0 1 0 1 0 1
0 0 1 1 0 0 1 1 0 0 1 1
(6.390)
(6.391)
AmC1 D
Am Am N Am Am
(6.392)
N where Am denotes the matrix that is obtained by taking the 1s complement of the elements of Am . A Walsh code of length 2m is obtained by taking the rows (or columns) of the Hadamard matrix Am and by mapping 0 into 1. From the construction of Hadamard matrices, it is easily seen that two words of a Walsh code are orthogonal. Figure 6.71 shows the 8 signals obtained with the Walsh code of length 8: the signals are obtained by interpolating the Walsh code sequences by a lter having impulse response wTc .t/ D rect t Tc =2 Tc (6.393)
Chapter 7
In this chapter we will reconsider amplitude modulation (PAM and QAM, see Chapter 6) for continuous transmission, taking into account the possibility that the transmission channel may distort the transmitted signal [1, 2]. We will also consider the effects of errors introduced by the digital transmission on a PCM encoded message (see Chapter 5) [3].
7.1
Let us briey examine the fundamental blocks of the baseband digital transmission system illustrated in Figure 7.1. The signals at various points of a ternary PAM system are shown in Figure 7.2. Source. We assume that a source, not explicitly indicated in Figure 7.1, generates a message fb` g composed of a sequence of binary symbols b` 2 f0; 1g, that are emitted every Tb seconds: fb` g D f: : : ; b
1 ; b0 ; b1 ; b2 ; : : : g
(7.1)
Usually fb` g is a sequence of i.i.d. symbols. The system bit rate, which is associated with the message fb` g, is equal to (see Section 6.2) 1 Rb D (bit/s) (7.2) Tb
Transmitter
Bit mapper. The bit mapper uses a one-to-one map to match a multilevel symbol to an input bit pattern. Let us consider, for example, symbols fak g from a quaternary alphabet, ak 2 A D f 3; 1; 1; 3g. To select the values of ak we consider pairs of input bits and map them into quaternary symbols as indicated in Table 7.1. Note that bits are mapped into symbols without introducing redundancy, therefore we speak of uncoded transmission, or in other words the sequence of symbols fak g is not obtained by applying channel coding.1 This situation will be maintained throughout the chapter.
1
We distinguish three types of coding: 1) source or entropy coding; 2) channel coding; and 3) line coding. Their objectives are respectively: 1) compress the digital message by lowering the bit rate without losing the original signal information (see Chapter 5); 2) increase the reliability of the transmission by inserting redundancy in the transmitted message, so that errors can be detected and/or corrected at the receiver (see Chapters 11 and 12); and 3) shape the spectrum of the transmitted signal (see Appendix 7.A).
540
Figure 7.2. Signals at various points of a ternary PAM transmission system with alphabet A D f 1; 0; 1g.
For uncoded quaternary transmission the symbol period or modulation interval T is given by T D 2Tb . 1=T is the modulation rate or symbol rate of the system and is measured in Baud: it indicates the number of symbols per second that are transmitted. In general, if the values of ak belong to an alphabet A with M elements, then 1 1 1 D (Baud) (7.3) T log2 M Tb
541
b2k 0 1 1 0
b2kC1 0 0 1 1
ak 3 1 1 3
We note that in Section 6.3 we considered an alphabet whose elements were indices, that is ak 2 f1; 2; : : : ; Mg. Now the values of ak are associated with fn g, n D 1; : : : ; M, that is ak 2 A D f .M 1/; : : : ; 1; 1; : : : ; .M 1/g. Modulator. For a PAM system, see (6.109), the modulator associates the symbol ak with the amplitude of a given pulse h T x : ak ! ak h T x .t kT / (7.4)
Therefore the modulated signal s.t/ that is input to the transmission channel is given by s.t/ D
C1 X kD 1
ak h T x .t
kT /
(7.5)
Transmission channel
The transmission channel is assumed to be linear and time invariant, with impulse response gCh . Therefore the desired signal at the output of the transmission channel still has a PAM structure. From the relation sCh .t/ D gCh s.t/ we dene qCh .t/ D h T x gCh .t/ then we have sCh .t/ D
C1 X kD 1
(7.6)
(7.7)
ak qCh .t
kT /
(7.8)
The transmission channel introduces an effective noise w. Therefore the signal at the input of the receive lter is given by: r.t/ D sCh .t/ C w.t/ (7.9)
542
Receiver
The receiver consists of three functional blocks: 1. Amplier-equalizer lter. This block is assumed linear and time invariant with impulse response g Rc . Then the desired signal is given by: s R .t/ D g Rc sCh .t/ Let the overall impulse response of the system be q R .t/ D qCh g Rc .t/ D h T x gCh g Rc .t/ then s R .t/ D In the presence of noise, r R .t/ D s R .t/ C w R .t/ where w R .t/ D w g Rc .t/. 2. Sampler. If the duration of q R .t/ is conned to a modulation interval, then the various pulses do not overlap and, in the absence of noise, sampling at instants t0 CkT yields:2 rk D r R .t0 C kT / D ak q R .t0 / D ak h 0 (7.14) (7.13)
C1 X kD 1
(7.10)
(7.11)
ak q R .t
kT /
(7.12)
where h 0 D q R .t0 / is the amplitude of the overall impulse response at the sampling instant t0 . The parameter t0 is called timing phase, and its choice is fundamental for system performance. 3. Threshold detector. From the sequence frk g we detect the transmitted sequence fak g. The simplest structure is the instantaneous non-linear threshold detector: ak D Q[rk ] O (7.15)
O where Q[rk ] is the quantizer characteristic with rk 2 < and ak 2 A, alphabet of ak . An example of quantizer characteristic for A D f 1; 0; 1g is given in Figure 7.3. From the sequence fak g, using an inverse bit mapper, the detected binary information O O message fb` g is obtained.
2
To simplify the notation, the sample index k, associated with the instant t0 C kT , here appears as a subscript.
543
^ = Q[r ] ak k h0 2 0 1 1
h0 2
rk
Figure 7.3. Characteristic of a threshold detector for ternary symbols with alphabet A D f 1; 0; 1g, and amplitude h0 of the overall impulse response.
We recall that the receiver structure described above was optimized in Chapter 6 for an ideal AWGN channel.
ak q.t
kT /
(7.16)
where q.t/ is the impulse response of a suitable lter. In other words, a PAM signal s.t/ may be regarded as a signal generated by an interpolator lter with impulse response q.t/, t 2 <, as shown in Figure 7.4. From the spectral analysis (see Example 1.9.9 on page 69) we know that s is a cyclostationary process with average power spectral density given by (see (1.398)): 2 1 N Ps . f / D Q. f / Pa . f / (7.17) T where Pa is the spectral density of the message and Q is the Fourier transform of q. From Figure 7.4 it is important to verify that by ltering s.t/ we obtain a signal that is still PAM, with a pulse given by the convolution of the lter impulse responses. The spectral density of a ltered PAM signal is obtained by multiplying Pa . f / in (7.17) by the squared magnitude of the lter frequency response. As Pa . f / is periodic of period 1=T , then the bandwidth B of the transmitted signal is equal to that of h T x .
ak T
Figure 7.4. The PAM signal as output of an interpolator lter.
s(t)
544
Occasionally the passband version of PAM that is obtained by DSB modulation is also proposed. Passband PAM, together with the associated parameters, is discussed in Appendix 7.C. Example 7.1.1 (PSD of an i.i.d. symbol sequence) Let fak g be a sequence of i.i.d. symbols with values from the alphabet A D f1 ; 2 ; : : : ; M g and p./, 2 A, be the probability distribution of each symbol. The mean value and the statistical power of the sequence are given by X X p./ Ma D jj2 p./ ma D
2A 2A
(7.18)
(7.19)
2 Consequently a D Ma jma j2 . Following Example 1.7.1 on page 51, that describes the decomposition of the PSD of a message into ordinary and impulse functions, the decomposition of the PSD of s is given by 2 2 2 .c/ N s . f / D 1 Q. f / a T D jQ. f /j2 a (7.20) P T T
` T
We note that spectral lines occur in the PSD of s if ma 6D 0 and Q. f / is non-zero at frequencies multiple of 1=T . Typically the presence of spectral lines is not desirable for a transmission scheme. We obtain ma D 0 by choosing an alphabet with symmetric values with respect to zero, and by assigning equal probabilities to antipodal values. In some applications, the spectrum of the transmitted signal is shaped by introducing correlation among transmitted symbols by a line encoder, see (7.17): in this case the sequence has memory. We refer the reader to Appendix 7.A for a description of the more common line codes.
7.2
Transmitter
Bit mapper. The bit mapper uses a map to associate a complex-valued symbol ak to an input bit pattern. In Figure 7.6 are given two examples of constellations and corresponding
545
ak,Q
(1000) (1100) (0100) (0000)
(1001)
(1101)
(0101)
(0001)
1 ak,Q (11)
(01) (10) (00)
-3
(1011)
-1
(1111)
1
(0111)
3 a (0011) k,I
-1 ak,I
(1010) (1110) (0110) (0010)
-3
(a) QPSK.
(b) 16-QAM.
and ak;I D Re [ak ] and ak;Q D Im [ak ]. 4-PSK (or QPSK) symbols are taken from an alphabet with four elements, each identied by two bits. Similarly each element in a 16QAM constellation is uniquely identied by four bits. Modulator. Typically the pulse h T x is real-valued, however, the baseband modulated signal is complex-valued: s .bb/ .t/ D D
C1 X kD 1 C1 X kD 1
ak h T x .t ak;I h T x .t
kT / kT / C j
C1 X kD 1
(7.24) ak;Q h T x .t kT /
546
-B
0 S (f)
f0 -B
f0
f 0 +B
-f0 -B
-f 0
-f 0+B
f0 -B
f0
f 0 +B
Let f 0 (!0 D 2 f 0 ) and '0 be, respectively, the carrier frequency (radian frequency) and phase. We dene s .C/ .t/ D 1 s .bb/ .t/e j .!0 tC'0 / 2
F
! S .C/ . f / D 1 S .bb/ . f 2
f 0 /e j'0
(7.25)
then the real-valued transmitted signal is given by: s.t/ D 2Refs .C/ .t/g
F
! S. f / D S .C/ . f / C S .C/ .
f/
(7.26)
The transformation in the frequency domain from s .bb/ to s is illustrated in Figure 7.7.
547
Moreover, starting from a relation similar to (1.304), we get that s is a cyclostationary random process with an average PSD given by3 N N Ps . f / D 1 [Ps .bb/ . f 4 N f 0 / C Ps .bb/ . f f 0 /] (7.28)
We note that the bandwidth B of the transmitted signal is equal to twice the bandwidth of h T x .
# kT /
(7.29)
The block-diagram representation of (7.29) is shown in Figure 7.8. As s .bb/ is in general a complex-valued signal, an implementation based on this representation requires a processor capable of complex arithmetic. 2. As e j .!0 tC'0 / D cos.!0 t C '0 / C j sin.!0 t C '0 / (7.30)
The result (7.28) needs clarication. We rst consider the situation where the condition rs .bb/ s .bb/ .t; t / D 0 is satised, as for example in the case of QAM with i.i.d. circularly symmetric symbols (see (1.407)). From the equation (similar to (1.304)) that relates rs to rs .bb/ and rs .bb/ s .bb/ , as the cross-correlations are zero, we nd that the process s is cyclostationary in t of period T . Taking the average correlation in a period T , the results (7.27) and (7.28) follow. We now consider the situation where rs .bb/ s .bb/ .t; t / 6D 0, and in particular the case where rs .bb/ s .bb/ .t; t / is a periodic function in t of period T , as for example in the case of PAM-DSB (see Appendix 7.C). In this situation the cross-correlations, in the equation similar to (1.304), do not vanish. If a real value T p exists, such that T p is an integer multiple of both T and 1= f 0 , then s is cyclostationary in t of period equal to T p . Taking the average correlation over the period T p , and expanding rs .bb/ s .bb/ .t; t / in Fourier series (in the variable t), for 1=T f 0 it happens that the autocorrelation terms approximate the same terms found in the previous case, and the cross-correlation terms become negligible.
548
ak;I h T x .t
kT /
sin.!0 t C '0 /
C1 X kD 1
ak;Q h T x .t
kT /
(7.31) The block-diagram representation of (7.31) is shown in Figure 7.5 (see also Figure 6.38). The implementation of a QAM transmitter based on (7.31) is discussed in Appendix 7.D. 3. Using the polar notation ak D jak je jk , (7.29) takes the form: " # C1 X s.t/ D Re e j .!0 tC'0 / jak je jk h T x .t kT / " D Re
C1 X kD 1 kD 1 C1 X kD 1
jak je
j .!0 tC'0 Ck /
# h T x .t kT / kT / (7.32)
If jak j is a constant we obtain the PSK signal (6.127), where the information bits select only the value of the carrier phase.
Coherent receiver
In the absence of noise the general scheme of a coherent receiver is shown in Figure 7.9, which follows the scheme of Figure 6.40. The received signal is given by: sCh .t/ D s gCh .t/
F
! SCh . f / D S. f /GCh . f /
(7.33)
First, the received signal is translated to baseband by a frequency shift, s M0 .t/ D sCh .t/e
j .!0 tC'1 /
! S M0 . f / D SCh . f C f 0 /e
j'1
(7.34)
! S R . f / D S M0 . f /G Rc . f /
(7.35)
549
G (f)
Ch
f0 S (f)
Ch
2f0
1
f0 B f0 G (f)
Rc
f0 +B
SMo(f)
2f0
SR (f)
2f0
Figure 7.10. Frequency responses of the channel and of signals at various points of the receiver.
.bb/ We note that, if g Rc is a non-distorting ideal lter with unit gain, then s R .t/ D .1=2/sCh .t/. In the particular case where g Rc is a real-valued lter, then the receiver in Figure 7.9 is simplied into that of Figure 7.5. Figure 7.10 illustrates these transformations. We note that in the above analysis, as the channel may introduce a phase offset, the receive carrier phase '1 may be different from the transmit carrier phase '0 .
7.3
Recalling the relations of Figure 1.30, we illustrate the baseband equivalent scheme with reference to Figure 7.11:4 by assuming that the transmit and receive carriers have the same
We note that the term e j'0 has been moved to the receiver; therefore the signals s .bb/ and r .bb/ of Figure 7.11 are dened apart from the term e j'0 . This is the same as assuming as reference carrier e j .2 f 0 tC'0 / .
550
frequency, we can study QAM systems by the same method that we have developed for PAM systems.
7.3.1
Signal analysis
We refer to Section 1.7.4 for an analysis of passband signals; we recall here that if for f > 0 the spectral density of w, Pw . f /, is an even function around the frequency f 0 , then the real and imaginary parts of w.bb/ .t/ D w I .t/ C jw Q .t/ have a power spectral density that is given by ( 2Pw . f C f 0 / f f0 1 (7.36) Pw I . f / D Pw Q . f / D Pw.bb/ . f / D 2 0 elsewhere Moreover, Pw I w Q . f / D 0, that is w I ? w Q , hence
2 2 2 2 w I D w Q D 1 w.bb/ D w 2
(7.37)
To simplify the analysis, for the study of a QAM system we will adopt the PAM model of Figure 7.1, assuming that all signals and lters are in general complex. We note that p the factor .1=2/e j .'1 '0 / appears in Figure 7.11. We will include the factor e j .'1 '0 / = 2 p in the impulse response of the transmission channel gCh , and the factor 1= 2 in the impulse response g Rc . Consequently the additive noise has a spectral density equal to .1=2/Pw.bb/ . f / D 2Pw . f C f 0 / for f f 0 . Therefore the scheme of Figure 7.1 holds also for QAM in the presence of additive noise: the only difference is that in the case of a QAM system the noise is complex-valued with orthogonal in-phase and quadrature components, each having spectral density Pw . f C f 0 / for f f 0 . Hence the scheme of Figure 7.12 is a reference scheme for both PAM and QAM, where 8 for PAM > GCh . f / < GC . f / D e j .'1 '0 / (7.38) > : GCh . f C f 0 /1. f C f 0 / for QAM p 2 We note that for QAM we have gC .t/ D e
j .'1 '0 /
p 2 2
(7.39)
551
Figure 7.12. Baseband equivalent model of PAM and QAM transmission systems.
With reference to the scheme of Figure 7.9, the relation between the impulse responses of the receive lters is given by g Rc .t/ D
1 p g Rc .t/ 2
(7.40)
In the following, to simplify the notation, the lter g Rc will be indicated in many passband schemes simply as g Rc . We summarize the denitions of the various signals in QAM systems. 1. Sequence of input symbols, fak g, sequence of symbols with values from a complexvalued alphabet A. In PAM systems, the symbols of the sequence fak g assume real values. 2. Modulated signal,5 s.t/ D 3. Signal at the channel output, sC .t/ D
C1 X kD 1 C1 X kD 1
ak h T x .t
kT /
(7.41)
ak qC .t
kT /
qC .t/ D h T x gC .t/
(7.42)
4. Circularly-symmetric, complex-valued, additive Gaussian noise, wC .t/ D w0I .t/ C jw0Q .t/ , with spectral density PwC . In the case of white noise it is:6 Pw0I . f / D Pw0Q . f / D and PwC . f / D N0 (V2 /Hz) In the model of PAM systems, only the component w0I is considered.
5 6
N0 (V2 /Hz) 2
(7.43)
(7.44)
We point out that for QAM s.t/ is in fact s .bb/ .t/. In fact (7.43) should include the condition f > f 0 . Because the bandwidth of g Rc is smaller than f 0 , this condition can be omitted.
552
5. Received or observed signal, rC .t/ D sC .t/ C wC .t/ 6. Signal at the output of the complex-valued amplier-equalizer lter g Rc , r R .t/ D s R .t/ C w R .t/ where s R .t/ D with q R .t/ D qC g Rc .t/ and w R .t/ D wC g Rc .t/ In PAM systems, g Rc is a real-valued lter. 7. Signal at the decision point at instant t0 C kT , yk D r R .t0 C kT / 8. Sequence of detected symbols, fak g O (7.50) (7.49) (7.48)
C1 X kD 1
(7.45)
(7.46)
ak q R .t
kT /
(7.47)
Signal-to-noise ratio
The performance of a system is expressed as a function of the signal-to-noise ratio 0 dened in (6.105), that we recall here. In general, with reference to the schemes of Figures 7.1 and 7.5, for a channel output signal, sCh , having minimum bandwidth Bmin , and assuming the noise w white with Pw . f / D N0 =2, we have 0D We express now E sCh
2 E[sCh .t/] MsCh E sCh D D .N0 =2/2Bmin N0 Bmin N0 .Bmin T / in the cases of PAM and QAM systems.
(7.51)
PAM systems. For an i.i.d. input symbol sequence, using (1.399) we have E sCh D Ma E qCh and, for Bmin D 1=.2T /, we obtain Ma E qCh N0 =2 where Ma is the statistical power of the data and E qCh is the energy of the pulse h T x gCh . Because for PAM, observing (7.38), we get qCh D qC , then (7.53) expressed as Ma E qC 0D N0 =2 0D (7.53) qCh D can be (7.54) (7.52)
553
(7.55)
Hence, as Bmin D 1=T , (7.51) becomes 0D We note that (7.56), expressed as7 0D Ma E qC =2 N0 =2 (7.57) Ma E qC N0 (7.56)
represents the ratio between the energy per component of sC , given by T E[.Re[sC .t/]/2 ] and T E[.Im[sC .t/]/2 ], and the PSD of the noise components, equal to N0 =2. Then (7.56), observing also (7.38), coincides with (7.53) of PAM systems.
7.3.2
Transmitter
The choice of the transmit pulse is quite important because it determines the bandwidth of the system (see (7.17) and (7.28)). Two choices are shown in Figure 7.13, where t T 2 1. h T x .t/ D wT .t/ D rect T , with wide spectrum; 2. h T x .t/ with longer duration and smaller bandwidth.
Transmission channel
The transmission channel is modelled as a time invariant linear system. Therefore it is represented by a lter having impulse response gCh . As described in Chapter 4, the majority of channels are characterized by frequency responses having a null at DC. Therefore the shape of the frequency response GCh . f / is as represented in Figure 7.14, where the passband goes from f 1 to f 2 . For transmission over cables, f 1 may be of the order of a few hundred Hertz, whereas for radio links, f 1 may be in the range of MHz or GHz. Consequently, PAM (possibly using a line code) as well as QAM transmission systems may be considered over cables; for transmission over radio, instead, a PAM signal needs to be translated in frequency (PAM-DSB or PAM-SSB), or a QAM system may be used, assuming as carrier frequency f 0 the center frequency of the passband ( f 1 ; f 2 ). In any case, the channel is bandlimited with a nite bandwidth f 2 f 1 .
7
The term Ma E qC =2 represents the energy of both Re[sC .t/] and Im[sC .t/], assuming that sC .t/ is circularly symmetric (see (1.407)).
554
With reference to the general model of Figure 7.12, we adopt the polar notation for GC : GC . f / D jGC . f /je j arg GC . f / (7.58)
Let B be the bandwidth of s.t/. According to (1.144), a channel presents ideal characteristics, known as Heaviside conditions for the absence of distortion, if the following two properties are satised: 1. the magnitude response is a constant for j f j < B, jGC . f /j D G0 for j f j < B (7.59)
2. the phase response is proportional to f for j f j < B, arg GC . f / D 2 f t0 for j f j < B (7.60)
Under these conditions, s is reproduced at the output of the channel without distortion, that is: sC .t/ D G0 s.t t0 / (7.61)
In practice, channels introduce both amplitude distortion and phase distortion. An example of frequency response of a radio channel is given in Figure 4.32: the overall effect is that the signal sC .t/ may be very different from s.t/. In short, for channels encountered in practice conditions (7.59) and (7.60) are too stringent; for PAM and QAM transmission systems we will refer instead to the Nyquist criterion (7.79).
555
Receiver
We return to the receiver structure of Figure 7.12, consisting of a lter g Rc followed by a sampler with sampling rate 1=T , and a data detector. In general, if the frequency response of the receive lter G Rc . f / contains a factor C.e j2 f T /, periodic of period 1=T , such that the following factorization holds: G Rc . f / D G M . f /C.e j2 f T / (7.62)
where G M . f / is a generic function, then the lter-sampler block before the data detector of Figure 7.12 can be represented as in Figure 7.15, where the sampler is followed by a discrete-time lter. It is easy to prove that in the two systems the relation between rC .t/ and yk is the same. Ideally, in the system of Figure 7.15 yk should be equal to ak . In practice, as illustrated in Figure 7.16, linear distortion and additive noise, the only disturbances considered here, may determine a signicant deviation of yk from the desired symbol ak .
556
4
2
2
1
1
k,Q
k,Q
y
1
1
3
2
3 3
y k,I
5 5
y k,I
(a) QPSK.
(b) 16-QAM.
Figure 7.16. Values of yk D yk,I C jyk,Q , in the presence of noise and linear distortion.
The last element in the receiver is the data detector. One of the simplest data detectors is the threshold detector, that associates with each value of yk a possible value of ak in the constellation. Using the rule of deciding for the symbol closest to the sample yk , the decision regions for a QPSK system and a 16-QAM system are illustrated in Figure 7.17.
7.3.3
Intersymbol interference
ai q R .t0 C .k
i/T /
(7.63)
557
yk,Q 3
(a) QPSK.
(b) 16-QAM.
Figure 7.17. Decision regions for a QPSK system and a 16-QAM system.
Introducing the version of q R , time-shifted by t0 , as h.t/ D q R .t0 C t/ and dening h i D h.i T / D q R .t0 C i T / it follows that s R;k D where ik D
C1 X i D 1; i 6Dk C1 X iD 1
(7.66)
(7.67)
ai h k
D ak h 0 C ik
(7.68)
ai h k
D C h
1 akC1
C h 1 ak
C h 2 ak
(7.69)
represents the intersymbol interference (ISI). The coefcients fh i gi 6D0 are called interferers. Moreover (7.65) becomes yk D ak h 0 C ik C w R;k (7.70)
We observe that, even in the absence of noise, the detection of ak from yk by a threshold detector takes place in the presence of the term ik , which behaves as a disturbance with respect to the desired term ak h 0 . For the analysis, it is often convenient to approximate ik as noise with a Gaussian distribution: the more numerous and similar in amplitude are the interferers, the more valid
558
is this approximation. In the case of i.i.d. symbols, the rst two moments of ik are easily determined. C1 X Mean value of ik : m i D ma hi (7.71)
i D 1; i 6D0 2 2 i D a C1 X i D 1; i 6D0
Variance of ik :
jh i j2
(7.72)
From (7.65), with fs R;k g given by (7.68), we derive the discrete-time equivalent scheme, with period T (see Figure 7.18), that relates the signal at the decision point to the data transmitted over a discrete-time channel with impulse response given by the sequence fh i g, called overall discrete-time equivalent impulse response of the system. Concerning the additive noise fw R;k g,8 being Pw R . f / D PwC . f /jG Rc . f /j2 the PSD of fw R;k g is given by Pw R;k . f / D
C1 X `D 1
(7.73) (7.74)
Pw R
1 T
In any case, the variance of w R;k is equal to that of w R and is given by Z C1 2 2 w R;k D w R D PwC . f /jG Rc . f /j2 d f
1
(7.75)
In particular, the variance per dimension of the noise is given by PAM QAM
2 I2 D E[w 2 ] D w R R;k 2 I2 D E[.Re[w R;k ]/2 ] D E[.I m[w R;k ]/2 ] D 1 w R 2
(7.76) (7.77)
In the case of PAM (QAM) transmission over a channel with white noise, where PwC . f / D N0 =2 (N0 ), (7.75) yields a variance per dimension equal to I2 D N0 E g Rc 2 (7.78)
559
where E g Rc is the energy of the receive lter. We observe that (7.78) holds for PAM as well as for QAM.
Nyquist pulses
The problem we wish to address consists in nding the conditions on the various lters of the system, so that, in the absence of noise, yk is a replica of ak . The solution is the Nyquist criterion for the absence of distortion in digital transmission. From (7.68), to obtain yk D ak it must be: Nyquist criterion in the time domain (
h0 D 1 hi D 0
i 6D 0
(7.79)
and ISI vanishes. A pulse h.t/ that satises the conditions (7.79) is said to be a Nyquist pulse with modulation interval T . The conditions (7.79) have their equivalent in the frequency domain. They can be derived using the Fourier transform of the sequence fh i g (1.90),
C1 X iD 1
hi e
j2 f i T
C1 1 X H f T `D 1
` T
(7.80)
where H. f / is the Fourier transform of h.t/. From the conditions (7.79) the left-hand side of (7.80) is equal to 1, hence the condition for the absence of ISI is formulated in the frequency domain for the generic pulse h as: Nyquist criterion in the frequency domain
C1 X `D 1
` T
DT (7.81)
From (7.81) we deduce an important fact: the Nyquist pulse with minimum bandwidth is given by: h.t/ D h 0 sinc t T
F
! H. f / D Th 0 rect
f 1=T
(7.82)
Denition 7.1 The frequency 1=.2T /, which coincides with half of the modulation frequency, is called Nyquist frequency. A family of Nyquist pulses widely used in telecommunications is composed of the raised cosine pulses whose time and frequency plots, for three values of the parameter , are illustrated in Figure 7.19a.
560
Figure 7.19. Time and frequency plots of raised cosine and square root raised cosine pulses for three values of the roll-off factor .
We dene 8 >1 > > > > > > > < 0 jxj 0 jxj 1 1 2 C A 1 2 jxj > 1 2 1C 2 (7.83)
rcos.x; / D
< jxj 1C 2
561
1 2
t C sinc T
1 2
1 t 2 2 T
(7.85)
It is easily proven that, from (7.83), the area of H. f / in (7.84) is equal to one, that is Z C1 f (7.86) ; df D 1 T rcos h.0/ D 1=T 1 and the energy is Z C1 f 2 2 Eh D ; df D T 1 T rcos 1=T 1 D H.0/h.0/ 1 4 4
(7.87)
We note that, from (7.84), the bandwidth of the baseband equivalent system is equal to .1 C /=.2T /. Consequently, for a QAM system the required bandwidth is .1 C /=T . Later we will also refer to square root raised cosine pulses, with frequency response given by s f (7.88) ; H. f / D T rcos 1=T and inverse Fourier transform t 1 t 1 t C cos C sinc C h.t/ D .1 / sinc .1 / T T 4 T 4 t 1 1 t C cos sinc T 4 T 4 t t t C 4 cos .1 C / sin .1 / T T T " D 2 # t t 1 4 T T In this case Z h.0/ D
1 C1
(7.89)
f ; df D 1 T rcos 1=T
(7.90)
C1 1
T 2 rcos
f ; 1=T
df D T
(7.91)
562
We note that H. f / in (7.88) is not the frequency response of a Nyquist pulse. Plots of h.t/ and H. f /, given respectively by (7.89) and (7.88), for various values of are shown in Figure 7.19b. The parameter , called excess bandwidth parameter or roll-off factor, is in the range between 0 and 1. We note that determines how fast the pulse decays in time. Observation 7.1 From the Nyquist conditions we deduce that: 1. a data sequence can be transmitted with modulation rate 1=T without errors if H. f / satises the Nyquist criterion and there is no noise; 2. the channel, with frequency response GC , must have a bandwidth equal to at least 1=.2T /, otherwise intersymbol interference cannot be avoided.
Eye diagram
From (7.68) we observe that if the samples fh i g, for i 6D 0, are not sufciently small with respect to h 0 , the ISI may result a dominant disturbance with respect to noise and impair the performance of the system. On the other hand, from (7.66) and (7.67) the discrete-time impulse response fh i g depends on the choice of the timing phase t0 (see Chapter 14) and on the pulse shape q R . In the absence of noise, at the decision point the sample y0 , as a function of t0 , is given by y0 D y.t0 / D
C1 X iD 1
ai q R .t0
iT/
(7.92)
ai q R .t0
1 q R .t0
iT/ T / C a2 q R .t0 2T / C
(7.93)
D Ca
C T / C a1 q R .t0
is the ISI. We now illustrate, through an example, a graphic method to represent the effect of the choice of t0 for a given pulse q R . We consider a PAM transmission system where y.t0 / is real: for a QAM system, both Re[y.t0 /] and Im[y.t0 /] need to be represented. We consider quaternary transmission with ak D n 2 A D f 3; 1; 1; 3g (7.94)
and pulse q R as shown in Figure 7.20. In the absence of ISI, i0 .t0 / D 0 and y0 D a0 q R .t0 /. In relation to each possible value n of a0 , the pattern of y0 as a function of t0 is shown in Figure 7.21: it is seen that the possible values of y0 , for n 2 A, are further apart, therefore they offer a greater
563
1.5
q (t)
0.5
0.5 T
T
t
2T
3T
4T
Figure 7.20. Pulse shape for the computation of the eye diagram.
n=3
2
n=1
nqR(t0)
=1 n
n=3
3
2T
3T
4T
t0
564
margin against noise in relation to the peak of q R , which in this example occurs at instant t0 D 1:5T . In fact, for a given t0 and for a given message : : : ; a 1 ; a1 ; a2 ; : : : , it may result in i0 .t0 / 6D 0, and this value is added to the desired sample a0 q R .t0 /. The range of variations of y0 .t0 / around the desired sample n q R .t0 / is determined by the values imax .t0 ; n / D imin .t0 ; n / D
fak g; a0 Dn fak g; a0 Dn
max min
i0 .t0 / i0 .t0 /
(7.95) (7.96)
The eye diagram is characterized by the 2M proles ( imax .t0 ; n / n q R .t0 / C n 2 A imin .t0 ; n /
(7.97)
If the symbols fak g are statistically independent with balanced values, that is both n and n belong to A, dening max D max n
n
(7.98)
and iabs .t0 / D max we have that imax .t0 / D iabs .t0 / imin .t0 / D iabs .t0 / (7.100) (7.101)
C1 X i D 1; i 6D0
jq R .t0
i T /j
(7.99)
We note that both functions do not depend on a0 D n . For the considered pulse, the eye diagram is given in Figure 7.22. We observe that as a result of the presence of ISI, the values of y0 may be very close to each other, and therefore reduce considerably the margin against noise. We also note that, in general, the timing phase that offers the largest margin against noise is not necessarily found in relation to the peak of q R . In this example, however, the choice t0 D 1:5T guarantees the largest margin against noise. In the general case, where there exists correlation between the symbols of the sequence fak g, it is easy to show that imax .t0 ; n / iabs .t0 / and imin .t0 ; n / iabs .t0 /. Consequently the eye may be wider as compared to the case of i.i.d. symbols. For quaternary transmission, we show in Figure 7.23 the eye diagram obtained with a raised cosine pulse q R , for two values of the roll-off factor. In general, the M 1 pupils of the eye diagram have a shape as illustrated in Figure 7.24, where two parameters are identied: the height a and the width b. The height a is an indicator of the noise immunity of the system. The width b indicates the immunity with respect to deviations from the optimum timing phase. For example a raised cosine pulse with D 1 offers greater immunity against errors in the choice of t0 as compared to the case D 0:125. The price we pay is a larger bandwidth of the transmission channel.
565
nqR(t0)+imax(t0;n)
nqR(t0)+imin(t0;n)
y0(t0)
T
t0
2T
3T
4T
Figure 7.22. Eye diagram for quaternary transmission and pulse qR of Figure 7.20.
We now illustrate an alternative method to obtain the eye diagram. A long random sequence of symbols fak g is transmitted over the channel, and the portions of the curve y.t/ D s R .t/ relative to the various intervals [t1 ; t1 C T /; [t1 C T; t1 C 2T /; [t1 C 2T; t1 C 3T /; : : : ], are mapped on the same interval, for example, on [t1 ; t1 C T /. Typically, we select t1 so that the center of the eye falls in the center of the interval [t1 ; t1 C T /. Then the contours of the obtained eye diagram correspond to the different proles (7.97). If the contours of the eye do not appear, it means that for all values of t0 the worst case ISI is larger than the desired component and the eye is shut. We note that, if the pulse q R .t/, t 2 <, has a duration equal to th , and dene Nh D dth =T e, we must omit plotting the values of y.t/ for the rst and last Nh 1 modulation intervals, as they would be affected by the transient behavior of the system. Moreover, for transmission with i.i.d. symbols, at every instant t 2 < the number of symbols fak g that contribute to y.t/ is at most equal to Nh . To plot the eye diagram we need in principle to generate all the M-ary symbol sequences of length Nh : in this manner we will reproduce the values of y.t/ in correspondence of the various proles.
7.3.4
Performance analysis
566
0.5
0.4
0.3
0.2
0.1
0 t/T
0.1
0.2
0.3
0.4
0.5
(a)
5
0.5
0.4
0.3
0.2
0.1
0 t/T
0.1
0.2
0.3
0.4
0.5
(b)
Figure 7.23. Eye diagram for quaternary transmission and raised cosine pulse qR with roll-off factor: (a) D 0:125 and (b) D 1.
a b t
0
567
For a memoryless decision rule on yk , i.e. regarding yk as an isolated sample, and still considering the ML detection criterion described in Section 6.1, we have the following correspondences: r R;k D yk D [yk;I ; yk;Q ] ak D [ak;I ; ak;Q ] w R;k D [Re[w R;k ]; Im[w R;k ]] ! r D [r1 ; r2 ]T ! s D [s1 ; s2 ]T ! w D [w1 ; w2 ]T (7.103) (7.104) (7.105)
If w R;k has a circularly symmetric Gaussian probability density function, as is the case if equation (7.43) holds, and the values assumed by ak are equally likely, then, given the observation yk , the detection criterion leads to choosing the value of n 2 A that is closest to yk .9 Moreover, the error probability depends on the distance dm between adjacent signals, in this specic case between adjacent values n 2 A, and on the variance per dimension of the noise w R;k , I2 . Hence, dening dm 2 (7.106) D 2 I and using (6.122) and (6.196), we have M-PAM Pe D 2 1 1 M Q. p
(7.107)
1 p Q. M-QAM / (7.108) p M We note that, for the purpose of computing Pe , only the variance of the noise w R;k is needed and not its PSD. We also note that (7.106) coincides with (6.57). With reference to Table 7.1 and to Figure 7.6, we consider dm D 2h 0 D 2. Now, for a channel with white noise, I2 is given by (7.78), hence from (7.106) it follows that Pe ' 4 1 D 2 N0 E g Rc (7.109)
Apparently, the above equation could lead to choosing a lter g Rc with very low energy, so that 1. However, here g Rc is not arbitrary, but it must be chosen such that the condition (7.79) for the absence of ISI is satised. We will see in Chapter 8 a criterion to design the lter g Rc . The general case of computation of Pe in the presence of ISI and non-Gaussian noise is given in Appendix 7.B.
We observe that this memoryless decision criterion is optimum only if the noise samples fw R;k g are statistically independent.
568
Assuming that the pulse qC that determines the signal at the channel output is given, the solution (see Section 1.10 on page 73) is provided by the receive lter g Rc matched to qC : hence the name matched lter (MF). In particular, with reference to the scheme of Figure 7.12 and for white noise wC , we have
G Rc . f / D K Q C . f / e j2 f t0
(7.110)
(7.111)
Substitution of (7.112) in (7.110) yields E g Rc D 1=E qC . Therefore (7.109) assumes the form D
MF
2E qC N0
(7.113)
The matched lter receiver is of interest also for another reason. Using (7.56) it is possible to determine the relation between the signal-to-noise ratios at the decision point and 0 at the receiver input: for a QAM system it turns out
MF
0
1 2 Ma
(7.114)
where Ma =2 is the statistical power per dimension of the symbol sequence. We stress the point that, for a certain modulation system with a given pulse qC and a given 0, it is not possible by varying the lter g Rc to obtain a higher at the decision point than (7.114), and consequently a better Pe . The equation (7.114) is often used as an upper bound of the system performance. However, we note that we have ignored the possible presence of ISI at the decision point that the choice of (7.110) might imply. We further observe that the matched lter receiver corresponds to the optimum receiver developed in Chapter 6. The only difference is that in that case the energy of the matched lter is equal to one. In Appendix 7.E we describe a Monte Carlo method for simulations of a discrete-time QAM system.
7.4
The carrierless AM/PM (CAP) modulation is a passband modulation technique that is closely related to QAM (see Section 7.2). The scheme of a QAM system is repeated for convenience in Figure 7.25. In CAP systems, the carrier is omitted by using passband lters and exploiting the periodicity of the PSD of the sequence fak g.
569
Using passband lters, the QAM scheme of Figure 7.25 is modied into the scheme of Figure 7.26, where the impulse responses of the transmit lters are given by h T x;I .t/ D h T x .t/ cos.2 f 0 t/ h T x;Q .t/ D h T x .t/ sin.2 f 0 t/ and the impulse responses of the receive lters are given by g Rc;I .t/ D g Rc .t/ cos.2 f 0 t/ g Rc;Q .t/ D g Rc .t/ sin.2 f 0 t/
. pb/ . pb/ . pb/ . pb/
(7.115) (7.116)
(7.117) (7.118)
Applying Theorem 1.1 we observe that, if f 0 is larger than the bandwidth of h T x , the . pb/ . pb/ hypothesis always veried in practice, then the pulses h T x;I and h T x;Q are related by the Hilbert transform (1.163); the same relation exists between the pulses g Rc;I and g Rc;Q . Consequently the two pulses (7.115) and (7.116) are orthogonal. Note that this property holds through the transmission channel. From (7.29), where for simplicity we set '0 D 0, in a QAM system the transmitted signal can be expressed as " # 1 X s Q AM .t/ D Re ak h T x .t kT / e j2 f 0 t
kD 1 1 X kD 1 . pb/ ak;I h T x;I .t Q . pb/ . pb/
(7.119) kT /
. pb/ ak;Q h T x;Q .t Q
kT /
570
where ak;I D Re[ak e j2 f 0 kT ] and ak;Q D Im[ak e j2 f 0 kT ]. Therefore in the scheme of Q Q Figure 7.26, the input to the modulation lters at instant k is given by ak D ak e j2 f 0 kT , Q which is equal to the symbol ak with an additional deterministic phase that must be removed at the receiver. CAP modulation is obtained by leaving out the additional phase, as shown in Figure 7.27. If we dene
. pb/ . pb/ Q h T x .t/ D h T x .t/e j2 f 0 t D h T x;I .t/ C j h T x;Q .t/
(7.120)
# Q ak h T x .t
. pb/
D
. pb/
ak;I h T x;I .t
. pb/
kT /
Because the pulses h T x;I .t/ and h T x;Q .t/, ltered by the transmission channel, are still related by the Hilbert transform, the receiver uses a passband matched lter of the phase-splitter type, implemented by two real-valued lters with impulse responses given by (7.117) and (7.118) (see Figure 7.27). In general, if the transfer function of the transmission medium is unknown a priori, the receive lters are adaptive (see Chapter 8). We note that CAP modulation is equivalent to QAM, with the difference that in a QAM Q system the input symbols fak g are substituted by the rotated symbols fak g. If f 0 is an integer multiple of 1=T , then there is no difference between CAP and QAM. QAM is usually selected if f 0 1=.2T /. In the case where f 0 is not much larger than 1=.2T /, as usually occurs in data transmission systems over metallic cables, CAP modulation may be preferred to QAM because it does not need carrier recovery. On the other hand, for transmission channels that introduce frequency offset, an acquisition mechanism of the carrier must be introduced and typically QAM is adopted.
571
7.5
This section is divided into two parts: the rst considers a PCM encoded signal (see Chapter 5) and evaluates the effects of digital channel errors on the reproduced analog signal, the second compares the performance of an analog transmission system with that of a digital system for the transmission of analog signals represented by the PCM method.
7.5.1
With reference to Figure 7.1 for PAM and to Figure 7.5 for QAM, the transformation that O maps input bits fb` g in output bits fb` g is called binary channel and is represented in Figure 7.28 (see also Figure 6.21 on page 457). A binary channel is typically characterized by the bit rate Rb D 1=Tb (bit/s) and by the bit error probability. The simplest model considers errors to be i.i.d., therefore the various distributions are obtainable starting from: O Pbit D P[b` 6D b` ] (7.122)
Correspondingly we give in Figure 7.29 the statistical model associated with a memoryless binary symmetric channel. In the following it is useful to evaluate the error probability of words c composed of b bits: Pe;c D 1 .1 Pbit /b (7.123) Pbit /b ' (7.124)
assuming errors are i.i.d.. On the other hand, if Pbit 1 it follows that .1 1 bPbit and Pe;c ' bPbit
1 bl 0 P bit
1 ^ bl 0
572
(7.125)
where, to simplify the analysis, we assume the rule 8 b 1 > < i D X c .k/2 j j jD0 > : Q i D sat C .i C 1 /1 2
(7.126)
In (7.125) c.k/ is the word of b bits transmitted over the binary channel, with components c j .k/ 2 f0; 1g, j D 0; 1; : : : ; b 1.
Figure 7.30. Composite transmission scheme of an analog signal via a binary channel.
573
We assume the binary channel symmetric and memoryless. Hence, if we express the generic detected bit at the output of the binary channel as c j .k/ 2 f0; 1g, the error probability Q is given by: P[c j .k/ 6D c j .k/] D Pbit Q (7.127)
as illustrated in Figure 7.29. Q If we denote with c.k/ D [cb 1 .k/; : : : ; c0 .k/] the received word, the bit mapper of the Q Q DAC performs the inverse operation of (7.125): c.k/ ! sq .kTc / D Q Q Q Q where 8 b 1 X > <D Q c j .k/2 j Q jD0 > : Q Q D sat C .Q C 1 /1 2 (7.128)
(7.129)
Given the one-to-one map between words and quantizer levels, using (7.124) it follows that P[Qq .kTc / 6D sq .kTc /] D Pe;c ' bPbit s (7.130)
where the two error terms are assumed uncorrelated, as they are to be ascribed to different phenomena. In particular, assuming the quantization noise uniform, from (5.41) it follows Meq D 2 12 D sat 12 3 22b (7.134)
574
The computation of MeCh is somewhat more difcult. First, from (7.126) and (7.132) we have b 1 X .c j .k/ c j .k//2 j Q (7.135) eCh .kTc / D 1
jD0
Let the error on the j-th transmitted bit be Q " j .k/ D c j .k/ then (7.135) becomes eCh .kTc / D 1
b 1 X jD0
c j .k/
(7.136)
" j .k/2 j
(7.137)
We note that " j .k/ 2 f 1; 0; 1g, with probabilities given by 8 > P[" j D 1] D 1 Pbit > 2 < > > : Then, observing (7.138), we get E[" j .k/] D 0 and E[" 2 .k/] D 1P[" j .k/ 6D 0] C 0P[" j .k/ D 0] j D P[c j .k/ 6D c j .k/] D Pbit Q For a memoryless binary channel E[" j1 .k/" j2 .k/] D hence from (7.137)
2 E[eCh .kTc /] D 12 Pbit b 1 X jD0
P[" j D
1] D
1 2
Pbit
(7.138)
P[" j D 0] D 1
Pbit (7.139)
(7.140)
E[" 21 ] D Pbit j 0
(7.141)
22 j D 12 Pbit
(7.142)
We note that, recalling footnote 3 on page 338, the statistical power of the output signal of an interpolator lter in a DAC is equal to the statistical power of the input samples. Consequently, from (7.133) the output signal-to-noise ratio is given by 3PCM D D D E[s 2 .t/] E[jQ .t/ s.t/j2 ] s E[s 2 .kTc /] E[jQq .kTc / s.kTc /j2 ] s Meq Ms C MeCh (7.143)
575
55
50
b=8
45
40
b=6
35
(dB)
30
PCM
25
b=4
20
15
b=2
10
0 8 10
10
10
10
10 P bit
10
10
10
10
Using (7.134) and (7.142), and for a signal-to-quantization noise ratio 3q D Ms =.12 =12/ (see (5.33)), we get 3PCM D 3q 1 C 4Pbit .22b 1/ (7.144)
We note that usually Pbit is such that Pbit 22b 1: thus it results 3PCM ' 3q , that is the output error is mainly due to the quantization error. In particular, for a signal s 2 U. sat ; sat ] whereby 3q D 22b , equation (7.144) is represented in Figure 7.31 for various values of b. For Pbit < 1=.4 22b / the output signal is corrupted mainly by the quantization noise, whereas for Pbit > 1=.4 22b / the output is affected mainly by errors introduced by the binary channel. For example for Pbit D 10 4 , going from b D 6 to b D 8 bits per sample yields an increment of 3PCM of only 2 dB. We observe that in the general case of non-uniform quantization there are no simple expressions similar to (7.142) and (7.144); however, the above observations remain valid.
7.5.2
Regenerative repeaters
The signal sent over a transmission line is attenuated and corrupted by noise. To cover long distances it is therefore necessary to place repeaters along the transmission line to restore the signal.
576
Analog transmission
The only solution possible in an analog transmission system is to place analog repeaters consisting of ampliers with suitable lters to restore the level of the signal and eliminate the noise outside the passband of the desired signal. The cascade of ampliers along a transmission line, however, deteriorates the signal-to-noise ratio. We consider the simplied scheme of Example 4.2.2 on page 271 with s.t/, transmitted signal with bandwidth B and available power Ps ; sCh .t/, desired signal at the output of transmission section i, with available power PsCh ; w.t/, effective noise at the input of repeater i; r.t/ D sCh .t/ C w.t/, overall signal at the amplier input of repeater i; s .t/, signal at the output of a system with N repeaters. Q We note that, if ac is the attenuation of the generic section i, then PsCh D 1 Ps ac (7.145)
In this example both the transmission channel and the various ampliers do not introduce distortion; the only disturbance in s .t/ is due to additive noise introduced by the various Q devices. For a source at noise temperature T0 , if F A is the noise gure of a single amplier, the signal-to-noise ratio at the amplier output of a single section is given by (4.92): 3D PsCh kT0 F A B (7.146)
Analogously for N analog repeater sections, as the overall noise gure is equal to F D N Fsr (see (4.77)), the overall signal-to-noise ratio, expressed as 3a D is given by 3a D 3 Ps D kT0 FB N (7.148) E[s 2 .t/] E[jQ .t/ s.t/j2 ] s (7.147)
Obviously in the derivation of (7.148) it is assumed that (4.83) holds, as a statistical power ratio is equated with an effective power ratio. Hence, in a system with analog repeaters, the noise builds up repeater after repeater and the overall signal-to-noise ratio worsens as the number of repeaters increases. Moreover, it must be remembered that in practical systems, possible distortion experienced by the desired signal through the various transmission channels and ampliers also accumulates, contributing to an increase of the disturbance in s .t/. Q
577
Digital transmission
In a digital transmission system, as an alternative to the simple amplication of the received signal r.t/, we can resort to the regeneration of the signal. With reference to the scheme of O Figure 7.32, given the signal r.t/, the digital message fb` g is rst reconstructed, and then re-transmitted by a modulator. Modeling each regenerative repeater by a memoryless binary symmetric channel (see Denition 6.1 on page 457) with error probability Pbit , and ignoring the probability that a bit undergoes more errors along the various repeaters, the bit error probability at the output of N regenerative repeaters is equal to10 Pbit;N ' 1 .1 Pbit / N ' N Pbit (7.149)
assuming Pbit 1, and errors of the different repeaters statistically independent. To obtain an expression of Pbit , it is necessary to specify the type of modulator. Let us consider an M-PAM system; then from (6.125) we get ! r 3 2.M 1/ Q Pbit D 0 (7.150) M log2 M M2 1 where from (6.108)11 PsCh (7.151) kT0 F A Bmin It is interesting to compare the bit error probability at the output of N repeaters in the two cases. r 1/ 3 0 (7.152) Analog repeaters: Pbit;N D 2.M M Q M log2 M2 1 N r 2.M 1/ N Q 3 0 Regenerative repeaters: Pbit;N D M log M (7.153) M2 1 2 0D Note that in (7.152) we used (7.148). Even if a regenerative repeater is much more complex than an analog repeater, for a given overall Pbit , regeneration allows a signicant saving in the power of the transmitted signal.
10 We note that a more accurate study shows that the errors have a Bernoulli distribution [4]. 11 To simplify the notation, we have indicated with the same symbol s Ch the desired signal at the amplier
input for both analog transmission and digital transmission. Note, however, that in the rst case sCh depends linearly on s, whereas in the second it represents the modulated signal that does not depend linearly on s.
578
Consequently, for an M-PAM modulator, the modulation interval T is equal to log2 M=Rb , and the minimum bandwidth of the transmission channel is equal to Bmin D b 1 D B 2T log2 M (7.155)
We note that the digital transmission of an analog signal may require a considerable expansion of the required bandwidth, if M is small. Obviously, using a more efcient digital representation of waveforms, for example by CELP, and/or a modulator with higher spectral efciency, for example, by resorting to multilevel transmission, Bmin may result very close to B or even smaller. Using (7.155) in (7.151), from (7.146) we have 0D log2 M 3 b (7.156)
The comparison between the two systems is based on the overall signal-to-noise ratio for the same transmitted power and transmission channel characteristics. To simplify the notation, initially we will consider a 2-PAM as modulator. Substituting the value of 0 given by (7.156) for M D 2 in (7.152) and (7.153), and recalling (7.144), valid for a uniform quantizer with 3q D 22b , that is assuming a uniform signal, see (5.44), we get 8 22b > > > q N analog repeaters > > > 3 > 1 C 4.22b 1/Q < bN (7.157) 3PCM D 2b 2 > > > q N regenerative repeaters > > > 3 > 1 C 4.22b 1/N Q : b Or else, using (7.148), we get 8 22b > > > q > > > 3a > 1 C 4.22b 1/Q < b > > > > > > > 1 C 4.22b : 22b 1/N Q q 3a N b N analog repeaters (7.158) N regenerative repeaters
3PCM D
579
45 b=7
40
b=6
35
b=5
30
(dB)
25
b=4
PCM
20
b=3
15
b=2
10
10
15
20
25
(dB)
30
35
40
45
Figure 7.33. 3PCM as a function of 3a for analog repeaters and 2-PAM. The parameter b denotes the number of bits for linear PCM representation.
In the case of analog repeaters, the plot of 3PCM as a function of 3a is given in Figure 7.33. We note that 3PCM is typically higher than 3a , as long as a sufciently large number of bits and 3a larger than 17 dB are considered. However, the PCM system is penalized by the increment of the bandwidth of the transmission channel. Using regenerative repeaters, for example N D 20 in Figure 7.34, 3PCM is always much higher than 3a , assuming an adequate number of bits for PCM coding is used. We note the threshold effect of Pbit as a function of 0 in a digital transmission system: if the ratio 0 is higher than a certain threshold, then Pbit is very small. Consequently, the quantization error becomes predominant at the receiver. While the previous graphs relate 3PCM directly to 3a , in practice it is interesting to determine the minimum value of 3 (or 0) so that 3PCM and 3a reach a certain value, say, of the order of 2040 dB, depending on the applications. We illustrate in Figure 7.35 these relations by varying the number N of repeaters and using a PCM encoder with b D 7. We show also a comparison for the same required bandwidth, which implies a modulator with M D 2b levels. In this case, with respect to 2-PAM, for the same Pbit the modulator requires an increment of about 6.b 1/ dB in terms of 0; therefore, from (7.156), the increment in terms of 3 is equal to 6.b 1/ 10 log10 .b 1/. The curve of 3PCM as a function of 3 for 128-PAM, plotted in Figure 7.35, is shifted to the right by about 28 dB with respect to 2-PAM. Therefore also for the same bandwidth, digital transmission is more efcient than analog transmission if the number of repeaters is large.
580
45 b=7
40
b=6
35
b=5
30
PCM (dB)
25
b=4
20
b=3
15
b=2
10
10
15
20
a (dB)
25
30
35
40
45
Figure 7.34. 3PCM as a function of 3a for 2-PAM transmission and N D 20 regenerative repeaters. The parameter b is the number of bits for linear PCM representation.
45
PCM
(N=10)
40
35
PCM
(N=100)
30
PCM , a (dB)
PCM(N=1000)
25
20
15
10
0 10
15
20
25
30
35 40 (dB)
45
50
55
60
65
Figure 7.35. 3a for analog transmission obtained by varying the number N of analog repeaters, and 3PCM for digital transmission with 2-PAM and b D 7, obtained by varying the number N of regenerative repeaters, as a function of 3 (signal-to-noise ratio of each repeater section). The dashed line represents 3PCM for 128-PAM and b D 7.
7. Bibliography
581
Figure 7.36. Minimum value of 3 as a function of the number N of regenerative repeaters required to guarantee an overall signal-to-noise ratio of 36 dB, for analog transmission and digital transmission with three different modulators. The number of bits for PCM coding is b D 7.
we illustrate in Figure 7.36 the minimum value of 3 as a function of the number of regenerative repeaters, for three different modulators.
Bibliography
[1] L. W. Couch, Digital and analog communication systems. Upper Saddle River, NJ: Prentice-Hall, 1997. [2] J. G. Proakis and M. Salehi, Communication system engineering. Englewood Cliffs, NJ: Prentice-Hall, 1994. [3] M. S. Roden, Analog and digital communication systems. Upper Saddle River, NJ: Prentice-Hall, 1996. [4] A. Papoulis, Probability, random variables and stochastic processes. New York: McGraw-Hill, 3rd ed., 1991.
582
[5] S. Benedetto and E. Biglieri, Principles of digital transmission with wireless applications. New York: Kluwer Academic Publishers, 1999. [6] P. Kabal and P. Pasupathy, Partial-response signaling, IEEE Trans. on Communications, vol. 23, pp. 921934, Sept. 1975. [7] D. L. Duttweiler, J. E. Mazo, and D. G. Messerschmitt, An upper bound on the error probability in decision-feedback equalization, IEEE Trans. on Information Theory, vol. 20, pp. 490497, July 1974. [8] G. Birkoff and S. MacLane, A survey of modern algebra. New York: Macmillan Publishing Company, 3rd ed., 1965. [9] D. G. Messerschmitt and E. A. Lee, Digital communication. Boston, MA: Kluwer Academic Publishers, 2nd ed., 1994. [10] B. R. Saltzberg, Intersymbol interference error bounds with application to ideal bandlimited signaling, IEEE Trans. on Information Theory, vol. 9, pp. 563568, July 1968. [11] R. Gitlin, J. Hayes, and S. Weinstein, Data communication principles. New York: Plenum Press, 1992. [12] M. C. Jeruchim, P. Balaban, and K. S. Shanmugan, Simulation of communication systems. New York: Plenum Press, 1992.
583
Appendix 7.A
The functions of line codes are: 1. to shape the spectrum of the transmitted signals, and match it to the characteristics of the channel (see (7.17)); this task may be performed also by the transmit lter; 2. to facilitate synchronization at the receiver, especially in case the information message contains long sequences of ones or zeros; 3. to improve system performance in terms of Pe . This appendix is divided in two parts: in the rst, several representations of binary symbols are listed; in the second, partial response systems are introduced. For in-depth study and analysis of spectral properties of line codes we refer to the bibliography, in particular [1, 5].
7.A.1
Line codes
With reference to Figure 7.37, the binary sequence fb` g, b` 2 f0; 1g, could be directly generated by a source, or be the output of a channel encoder. The sequence fak g is produced by a line encoder. The channel input is a PAM signal s.t/, obtained by modulating a rectangular pulse h T x .
584
NRZL
2
NRZM
2
6 t/T
10
2 0
6 t/T
10
NRZS
2
Dicode NRZ
2
6 t/T
10
2 0
6 t/T
10
585
Unipolar RZ
2
1.5
Polar RZ
2
1
1
0.5
0
1
0.5
1
0
6 t/T
10
2 0
6 t/T
10
Bipolar RZ
2
Dicode RZ
2
6 t/T
10
2 0
6 t/T
10
do not create synchronization problems. It is easy to see, however, that this line code leads to a doubling of the transmission bandwidth. 2. Biphase mark (B- -M) or Manchester 1: A transition occurs at the beginning of every bit interval; 1 is represented by a second transition within the bit interval, 0 is represented by a constant level. 3. Biphase space (B- -S): A transition occurs at the beginning of every bit interval; 0 is represented by a second transition within the bit interval, 1 is represented by a constant level. Biphase line codes are illustrated in Figure 7.40.
586
BiphaseL
2
BiphaseM
2
6 t/T
10
2 0
6 t/T
10
BiphaseS
2
Delay Modulation
2
6 t/T
10
2 0
6 t/T
10
The KBNT codes are an example of block line codes where the output symbol alphabet is ternary f 1; 0; 1g.
with
bk 2 f0; 1g
(7.161)
At the decoder the bits of the information sequence may be recovered by O O bk D a k C bk O Note that ak 2 f 1; 0; 1g; in particular ( ak D
1
(7.162)
1 0
if bk 6D bk if bk D bk
1 1
(7.163)
From (7.161), the relation between the PSDs of the sequences fak g and fbk g is given by Pa . f / D Pb . f / j1 e
j2 f T 2
D Pb . f / 4 sin2 . f T /
587
Therefore Pa . f / exhibits zeros at frequencies that are integer multiples of 1=T , in particular at f D 0. Moreover, from (7.161) we have ma D 0, independently of the distribution of fbk g. If the power of the transmitted signals is constrained, a disadvantage of the encoding method (7.161) is a reduced noise immunity with respect to antipodal transmission, that is for ak 2 f 1; 1g, because a detector at the receiver must now decide among three levels. O Moreover, long sequences of information bits fbk g that are all equal to 1 or 0 generate sequences of symbols fak g that are all equal: this is not desirable for synchronization. In any case, the biggest problem is the error propagation at the decoder, which, observing O (7.162), given that an error occurs in fak g, generates a sequence of bits fbk g that are in O error until another error occurs in fak g. This problem can be solved by precoding: from the O sequence of bits fbk g we rst generate the sequence of bits fck g, with ck 2 f0; 1g, by c k D bk c k where denotes the modulo 2 sum. Next, a k D ck with ak 2 f 1; 0; 1g. Hence, it results in ( 1 ak D 0 ck
1 1
(7.164)
(7.165)
if bk D 1 if bk D 0
(7.166)
In other words, a bit bk D 0 is mapped into the symbol ak D 0, and a bit bk D 1 is mapped alternately in ak D C1 or ak D 1. Consequently, from (7.166) decoding may be performed simply by taking the magnitude of the detected symbol: O O bk D jak j (7.167)
It is easy to prove that for a message fbk g with statistically independent symbols, and p D P[bk D 1], we have sin2 . f T / Pa e j2 f T D 2 p.1 p/ (7.168) p2 C .1 2 p/ sin2 f T The plot of Pa e j2 f T is shown in Figure 7.41 for different values of p. Note that the PSD presents a zero at f D 0. Also in this case ma D 0. We observe that the AMI line code is a particular case of the partial response system named dicode [6].
7.A.2
From Section 7.1, we recall in Figure 7.42 the block diagram of a baseband transmission system, where the symbols fak g belong to the following alphabet12 of cardinality M: ak 2 A D f .M 1/; .M 3/; : : : ; .M 3/; .M 1/g (7.169) and w.t/ is an additive white Gaussian noise.
12 In the present analysis only M-PAM systems are considered; for M-QAM systems the results can be extended
588
ak T
h Tx
s(t)
sCh (t)
r (t) Ch w(t)
Ch
rR (t)
yk t 0 +kT
^ ak
Rc
We assume that the transmission channel is ideal: the overall system can then be represented as an interpolator lter having impulse response q.t/ D h T x g Rc .t/ (7.170)
A noise signal w R .t/, obtained by ltering w.t/ by the receive lter, is added to the desired signal. Sampling the received signal at instants t0 CkT yields the sequence fyk g, as illustrated in Figure 7.43a. The discrete-time equivalent of the system is shown in Figure 7.43b, where fh i D q.t0 C i T /g, and w R;k D w R .t0 C kT /. We assume that fh i g is equal to zero for i < 0 and i N . The partial response (PR) polynomial of the system is dened as l.D/ D
N 1 X i D0
li D i
(7.171)
where the coefcients fli g are equal to the samples fh i g, and D is the unit delay operator.
589
ak T
l(D)
ak
(t)
rR (t)
yk t 0 +kT
^ ak
w R (t)
Figure 7.44. PR version of the system of Figure 7.42.
A PR system is illustrated in Figure 7.44, where l.D/ is dened in (7.171), and g is an analog lter satisfying the Nyquist criterion for the absence of ISI,
C1 X mD 1
G f
m DT T
(7.172)
The symbols at the output of the lter l.D/ in Figure 7.44 are given by
.t/ ak D N 1 X i D0
li ak
(7.173)
Note that the overall scheme of Figure 7.44 is equivalent to that of Figure 7.43a with q.t/ D
N 1 X i D0
li g.t
iT/
(7.174)
Also, observing (7.172), the equivalent discrete-time model is obtained for h i D li . In other words, from (7.174) the system of Figure 7.42 is decomposed into two parts: a lter with frequency response l.e j2 f T /, periodic of period 1=T , that forces the system to have an overall discrete-time impulse response equal to fh i g; an analog lter g that does not modify the overall lter h.D/ and limits the system bandwidth. As it will be clear from the analysis, the decomposition of Figure 7.44, on one hand, allows simplication of the study of the properties of the lter h.D/, and, on the other, to design an efcient receiver. The scheme of Figure 7.44 suggests two possible ways to implement the system of Figure 7.42:
590
ak T
l(D)
ak(t)
h Tx
(PR)
s(t)
Ch
g (PR)
Rc
rR (t)
yk t 0 +kT
^ ak
w(t)
Figure 7.45. Implementation of a PR system using a digital lter.
1. Analog: the system is implemented in analog form; therefore the transmit lter h T x and the receive lter g Rc must satisfy the relation HT x . f / G Rc . f / D Q. f / D l.e
j2 f T
/ G. f /
(7.175)
2. Digital: the lter l.D/ is implemented as a component of the transmitter by a digital lter; then the transmit lter h .P R/ and receive lter g .P R/ must satisfy the relation Tx Rc H.P R/ . f / G .P R/ . f / D G. f / Tx Rc (7.176)
The implementation of a PR system using a digital lter is shown in Figure 7.45. Note from (7.172) that in both relations (7.175) and (7.176) g is a Nyquist lter.
/ G. f / D 0
jfj>
Correspondingly, observing (7.174) the lter q assumes the expression N 1 X t iT q.t/ D li sinc T i D0
(7.179)
b) Spectral zeros at f D 1=.2T /. From the theory of signals, it is known that if Q. f / and its rst .n 1/ derivatives are continuous and the n-th derivative is discontinuous, then jq.t/j asymptotically decays as 1=jtjnC1 . The continuity of Q. f / and of its derivatives helps to reduce the portion of energy contained in the tails of q.t/. It is easily proven that in a minimum bandwidth system, the .n 1/-th derivative of Q. f / is continuous if and only if l.D/ has .1 C D/n as a factor. On the other hand, if l.D/ has a zero of multiplicity greater than one in D D 1, then the transition band of G. f / around f D 1=.2T / can be widened, thus simplifying the design of the analog lters.
591
c) Spectral zeros at f D 0. A transmitted signal with attenuated spectral components at low frequencies is desirable in many cases, e.g. for the implementation of SSB modulators (see Example 1.7.4 on page 58), or for transmission over channels with frequency responses that exhibit a spectral null at the frequency f D 0. Note that a zero of l.D/ in D D 1 corresponds to a zero of l.e j2 f T / at f D 0. d) Number of output levels. From (7.173), the symbols at the output of the lter l.D/ have an alphabet A.t/ of cardinality M .t/ . If we indicate with n l the number of coefcients of l.D/ different from zero, then the following inequality for M .t/ holds n l .M 1/ C 1 M .t/ M n l (7.180) In particular, if the coefcients fli g are all equal, then M .t/ D n l .M 1/ C 1. We note that, if l.D/ contains more than one factor .1 D/, then n l increases and, observing (7.180), also the number of output levels increases. If the power of the transmitted .t/ signal is constrained, detection of the sequence fak g by a threshold detector will cause a loss in system performance. e) Some examples of minimum bandwidth systems. In the case of minimum bandwidth systems, it is possible to evaluate the expression of Q. f / and q.t/ once the polynomial l.D/ has been selected. As the coefcients fli g are generally symmetric or antisymmetric around i D .N 1/=2, it is convenient to consider the time-shifted pulse .N 1/T q.t/ D q t Q 2 (7.181) Q Q. f / D e j f .N
1/T
Q. f /
In Table 7.2 the more common polynomials l.D/ are described, as well as the correQ sponding expressions of Q. f / and q.t/, and the cardinality M .t/ of the output alphabet A.t/ . Q In the next three examples, polynomials l.D/ that are often found in practical applications of PR systems are considered. Example 7.A.1 (Dicode lter) The dicode lter introduces a zero at frequency f D 0 and has the following expression l.D/ D 1 D
j2 f T ,
/ D 2j e
j f T
sin. f T /
Example 7.A.2 (Duobinary lter) The duobinary lter introduces a zero at frequency f D 1=.2T / and has the following expression l.D/ D 1 C D (7.184)
592
1C D 1 D D2
2T cos. f T / j 2T sin. f T / j 2T sin.2 f T / 4T cos2 . f T / D3 j 4T cos. f T / sin.2 f T / 4T sin. f T / sin.2 f T / 4T sin2 .2 f T / T C T cos.2 f T / C j 3T sin.2 f T / T C T cos.4 f T / C j 3T sin.4 f T /
2M 2M
1 1
2M
1 C 2D C D 2 D2
4M
1C D
4M
D2 C D3
4M
2D 2 C D 4 D2 D4
4M
2C D D2
4M
4M
The frequency response is given by l.e Observing (7.179) we have q.t/ D sinc
j2 f T
/ D 2e
j f T
cos. f T /
(7.185)
t T t C sinc T T
(7.186)
The plot of the impulse response of a duobinary lter is shown in Figure 7.46 with a continuous line. We notice that the tails of the two sinc functions cancel each other, in line with what was stated at point b) regarding the aymptotical decay of the pulse of a PR system with a zero in D D 1. Example 7.A.3 (Modied duobinary lter) The modied duobinary lter combines the characteristics of duobinary and dicode lters, and has the following expression l.D/ D .1 The frequency response becomes l.e
j2 f T
D/ .1 C D/ D 1
j4 f T
D2 sin.2 f T /
(7.187)
/D1
D 2j e
j2 f T
(7.188)
593
1.5
0.5
q(t)
0
0.5
1
1 t/T
Figure 7.46. Plot of q.t/ for duobinary ( ) and modied duobinary (- -) lters.
The plot of the impulse response of a modied duobinary lter is shown in Figure 7.46 with a dashed line.
f) Transmitted signal spectrum. With reference to the PR system of Figure 7.45, the spectrum of the transmitted signal is given by (see (7.17)) 2 1 l.e j2 f T / H.P R/ . f / Pa . f / Ps . f / D (7.190) Tx T For a minimum bandwidth system, with H.P R/ . f / given by (7.178), (7.190) simplies into Tx 8 1 > jl.e j2 f T /j2 P . f / < jfj a 2T Ps . f / D (7.191) 1 > :0 jfj > 2T In Figure 7.47 the PSD of a minimum bandwidth PR system is compared with that of a PAM system. The spectrum of the sequence of symbols fak g is assumed white. For the PR system, a modied duobinary lter is considered, so that the spectrum is obtained as the
594
product of the functions jl. e j2 f T /j2 D j2 sin.2 f T /j2 and jH.P R/ . f /j2 D T 2 rect. f T /, Tx plotted with continuous lines. For the PAM system, the transmit lter h T x is a square root raised cosine with roll-off factor D 0:5, and the spectrum is plotted with a dashed line.
li ak
(7.192)
The term l0 ak is the desired part of the signal s R;k , whereas the summation represents the ISI term that is often designated as controlled ISI, as it is deliberately introduced. .t/ The receiver detects the symbols fak g using the sequence of samples fyk D ak C w R;k g. 13 We discuss four possible solutions. 1. LE-ZF. A zero-forcing linear equalizer (LE-ZF) having D transform equal to 1=l.D/ is used. At the equalizer output, at instant k the symbol ak plus a noise term is
13 For a rst reading it is suggested that only solution 3 is considered. The study of the other solutions should
595
Figure 7.48. Four possible solutions to the detection problem in the presence of controlled ISI.
obtained; the detected symbols fak g are obtained by an M-level threshold detector, O as illustrated in Figure 7.48a. We note, however, that the amplication of noise by the lter 1=l.D/ is innite at frequencies f such that l.e j2 f T / D 0. 2. DFE. A second solution resorts to a decision-feedback equalizer (DFE), as shown in Figure 7.48b. An M-level threshold detector is also employed by the DFE, but there is no noise amplication as the ISI is removed by the feedback lter, having D transform equal to 1 l.D/=l0 . We observe that at the decision point the signal yk has the expression N 1 yk D Q l0 If we indicate with ek D ak we obtain
.t/ ak
C w R;k
N 1 X i D1
! li ak O
i
(7.193)
1 yk D ak C Q l0
! li ek
i
w R;k C
(7.194)
The equation (7.194) shows that a wrong decision negatively inuence successive decisions: this phenomenon is known as error propagation. 3. Threshold detector with M .t/ levels. This solution, shown in Figure 7.48c, exploits .t/ the M .t/ -ary nature of the symbols ak , and makes use of a threshold detector with .t/ levels followed by a LE-ZF. This structure does not lead to noise amplication M as solution 1, because the noise is eliminated by the threshold detector; however, there is still the problem of error propagation. 4. Viterbi algorithm. This solution, shown in Figure 7.48d, corresponds to maximumlikelihood sequence detection (MLSD) of fak g. It yields the best performance.
596
Solution 2 using the DFE is often adopted in practice: in fact it avoids noise amplication and is simpler to implement than the Viterbi algorithm. However, the problem of error propagation remains. In this case, using (7.194) the error probability can be written as # " N 1 X 1 P w R;k C Pe D 1 li ek i > l0 (7.195) M i D1 A lower bound Pe;L can be computed for Pe by assuming the error propagation is absent, or setting fek g D 0, 8k, in (7.195). If we denote by w R the standard deviation of the noise w R;k , we obtain l0 1 (7.196) Q Pe;L D 2 1 M w R Assuming w R;k white noise, an upper bound Pe;U is given in [7] in terms of Pe;L : Pe;U D .M=.M M N 1 Pe;L 1// Pe;L .M N
1
1/ C 1
(7.197)
From (7.197) we observe that the effect of the error propagation is that of increasing the error probability by a factor M N 1 with respect to Pe;L . A solution to the problem of error propagation is represented by precoding, which will be investigated in depth in Chapter 13.
Precoding
We make use here of the following two simplications: 1. the coefcients fli g are integer numbers; 2. the symbols fak g belong to the alphabet A D f0; 1; : : : ; M because arithmetic modulo M is employed. We dene the sequence of precoded symbols fak g as: N ! N 1 X . p/ . p/ li ak i mod M N a k l0 D a k N
i D1 . p/
(7.198)
We note that (7.198) has only one solution if and only if l0 and M are relatively prime [8]. In case l0 D D l j 1 D 0 mod M, and l j and M are relatively prime, (7.198) becomes ! N 1 X . p/ . p/ ak j l j D ak N li ak i mod M N (7.199)
i D jC1
597
. p/
li ak N
. p/ i
(7.200)
From the comparison between (7.198) and (7.200), or in general (7.199), we have the fundamental relation
.t/ ak mod M D ak
(7.201)
.t/ Equation (7.201) shows that, as in the absence of noise we have yk D ak , the symbol ak can be detected by considering the received signal yk modulo M; this operation is O memoryless, therefore the detection of ak is independent of the previous detections fak i g, O i D 1; : : : ; N 1. Therefore the problem of error propagation is solved. Moreover, the desired signal is not affected by ISI. If the instantaneous transformation
ak
. p/
. p/
D 2ak N
. p/
.M
1/
(7.202)
is applied to the symbols fak g, then we obtain a sequence of symbols that belong to the N . p/ in (7.169). The sequence fa . p/ g is then input to the lter l.D/. Precoding alphabet A k consists of the operation (7.198) followed by the transformation (7.202). However, we note that (7.201) is no longer valid. From (7.202), (7.200), and (7.198), we obtain the new decoding operation, given by ! .t/ ak ak D C K mod M (7.203) 2 where K D .M 1/
N 1 X i D0
li
(7.204)
A PR system with precoding is illustrated in Figure 7.49. The receiver is constituted by a threshold detector with M .t/ levels that provides the symbols fak g, followed by a block O .t/ that realizes (7.203) and yields the detected data fak g. O
a (t) l(D)
k
yk
^ (t) ak
decoder
^ ak
598
If we assume that the cardinality of the set A.t/ is maximum, i.e. M .t/ D M n l , then the .t/ output levels are equally spaced and the symbols ak result equally likely with probability
.t/ P[ak D ] D
1 M nl
2 A.t/
(7.205)
.t/ In general, however, the symbols fak g are not equiprobable, because several output levels are redundant, as can be deduced from the following example.
Example 7.A.4 (Dicode lter) We assume M D 2, therefore ak D f0; 1g; the precoding law (7.198) is simply an exclusive or and ak N
. p/ . p/
D ak ak N
. p/ 1
(7.206)
The symbols fak g are obtained from (7.202), ak they are antipodal as ak are given by
. p/ . p/
D 2ak N
. p/
(7.207)
ak
D 2 ak N
ak N
(7.208)
.t/ The values of ak 1 , ak , ak and ak are given in Table 7.3. We observe that both output N N levels 2 correspond to the symbol ak D 1 and therefore are redundant; the three levels are not equally likely. The symbol probabilities are given by .t/ P[ak D 2] D .t/ P[ak 1 4 1 2
. p/
. p/
(7.209)
D 0] D
Figure 7.50a shows the precoder that realizes equations (7.206) and (7.207). The decoder, O realized as a map that associates the symbol ak D 1 to 2, and the symbol ak D 0 to 0, is O illustrated in Figure 7.50b.
Table 7.3 Precoding for the dicode lter.
ak N
. p/ 1
ak 0 1 0 1
ak N
. p/
.t/ ak
0 0 1 1
0 1 1 0
0 C2 0 2
599
0 1
D
(a) precoder
^ (t) a
k
0 2
0 1
^ ak
(b) decoder
Figure 7.50. Precoder and decoder for a dicode lter l.D/ with M D 2.
Pbit D Q
1 I
(7.211)
ak
ak
(p)
(p) -1 a k +1
(7.210)
600
.t/ b) Duobinary signal with precoding. The transmitted signal is now given by ak D ak C . p/ . p/ ak 1 2 f 2; 0; 2g, where ak 2 f 1; 1g is given by (7.202) and (7.198). The received signal is given by
. p/
yk D ak C w B;k Q
2 where the variance of the noise is w B D 2 I2 , as 2.t/ D 2. Q ak
.t/
(7.212)
At the receiver, using a threshold detector with thresholds set at 1, we have the following conditional error probabilities: 1 .t/ P[E j ak D 0] D 2Q w B Q 1 .t/ .t/ P[E j ak D 2] D P[E j ak D 2] D Q w B Q Consequently, at the detector output we have Pbit D P[ak 6D ak ] O
.t/ .t/ D P[E j ak D 0] 1 C P[E j ak D 2] 1 2 2 1 D 2Q p 2 I
We observe a worsening of about 3 dB in terms of the signal-to-noise ratio with respect to case a). c) Duobinary signal. given by
.t/ The transmitted signal is ak D ak C ak 1.
yk D ak C ak
C wC;k Q
2 where wC D 2 I2 . We consider using a receiver that applies MLSD to recover the data; Q from Example 8.12.1 on page 687 it results in p ! 8 1 (7.214) Pbit D K Q DKQ 2wC I Q
where K is a constant. We note that the PR system employing MLSD at the receiver achieves a performance similar to that of a system transmitting antipodal signals, as MLSD exploits the correlation .t/ between symbols of the sequence fak g. Example 7.A.6 (Equivalent channel g of the type 1 C D) In this example it is the channel itself that forms a duobinary signal.
601
C w D;k Q
(7.215)
2 where w D D 2 I2 . Q An attempt at pre-equalizing the signal at the transmitter by inserting a lter l.D/ D .t/ 1=.1 C D/ D 1 D C D 2 C would yield symbols ak with unlimited amplitude; therefore such a conguration cannot be used. Equalization at the receiver using the scheme of Figure 7.48a would require a lter of the type 1=.1 C D/, which would lead to unlimited noise enhancement. Therefore we resort to the scheme of Figure 7.48c, where the threshold detector has thresholds set at 1. To avoid error propagation, we precode the message and transmit the . p/ sequence fak g instead of fak g. At the receiver we have
yk D ak
. p/
C ak
. p/ 1
C w D;k Q
(7.216)
(7.217)
e) MLSD receiver. To detect the sequence of information bits from the received signal (7.215), MLSD can be adopted. Pbit is in this case given by (7.214).
602
Appendix 7.B
7.B.1
In the absence of ISI, the signal at the decision point is the type (7.102) (7.218)
where w R;k is the sample of an additive noise signal. Assuming fw R;k g stationary with probability density function pw . /, from (7.218) for ak D n 2 A we have p yk jak . j n / D pw . Therefore the MAP criterion (6.26) becomes 2 Rm a k D m O if m D arg max pn pw .
n
h 0 n /
(7.219)
h 0 n /
(7.220)
We consider now the application of the MAP criterion to an M-PAM system, where n D 2n 1 M n D 1; : : : ; M (7.221)
The decision regions fRn g, n D 1; : : : ; M, are formed by intervals, or, in general, by the union of intervals, whose boundary points are called decision thresholds fi g, i D 1; : : : ; M 1. Example 7.B.1 (Determination of the optimum decision threholds) We consider a 4-PAM system with the following symbol probabilities: 3 3 1 1 ; ; ; f p1 ; p2 ; p3 ; p4 g D 20 20 2 5 The noise is assumed to have an exponential probability density function pw . / D e 2
j j
(7.222)
(7.223)
pn pw .
h 0 n /
n D 1; : : : ; 4
(7.224)
are illustrated in Figure 7.51. We note that, for the choice in (7.222) of the symbols probabilities, the decision thresholds, also shown in Figure 7.51, are obtained from the intersections between curves in (7.224) relative to two adjacent symbols; therefore they are given by the solutions of the M 1 equations pi pw .i h 0 i / D pi C1 pw .i h 0 i C1 / i D 1; : : : ; M 1 (7.225)
603
Figure 7.51. Optimum thresholds for a 4-PAM system with non-equally likely symbols.
We point out that, if the probability that the symbol ` is sent is very small, p` 1, the measure of the corresponding decision interval could be equal to zero, and consequently this symbol would never be detected. In this case the decision thresholds will be fewer than M 1. Example 7.B.2 (Computation of Pe for a 4-PAM system) We indicate with Fw .x/ the probability distribution of w R;k : Z x pw . / d Fw .x/ D
1
(7.226)
For a M-PAM system with thresholds 1 ; 2 , and 3 , the probability of correct decision is given by (6.18): 4 X Z pn pw . h 0 n / d P[C] D
nD1 Rn
1 1
D p1 Z C p3
pw . pw .
h 0 1 / d C p2
2 1
pw .
C1
h 0 2 / d (7.227) h 0 4 / d h 0 2 /] Fw .3 h 0 4 /]
3 2
h 0 3 / d C p4
pw .
D p1 [Fw .1 C p3 [Fw .3
h 0 1 /] C p2 [Fw .2 h 0 3 / Fw .2
h 0 2 /
Fw .1
h 0 3 /] C p4 [1
604
We note that, if Fw is a continuous function, optimum thresholds can be obtained by equating to zero the derivative of the expression in (7.227) with respect to 1 ; 2 , and 3 . In the case of equally likely symbols and equidistant thresholds, i.e. i D h 0 .2i equation (7.227) yields P[C] D 1 2 1 1 M Fw . h 0 / (7.229) M/ i D 1; : : : ; M 1 (7.228)
We note that (7.229) is in agreement with (6.122) obtained for Gaussian noise.
7.B.2
We consider M-PAM transmission in the presence of ISI. We assume the symbols in (7.221) are equally likely and the decision thresholds are of the type given by (7.228). With reference to (7.65), the received signal at the decision point assumes the following expression: yk D h 0 ak C ik C w R;k where ik represents the ISI and is given by X ik D h i ak
i 6D0
(7.230)
(7.231)
and w R;k is Gaussian noise with statistical power 2 and statistically independent of the i.i.d. symbols of the message fak g. We examine various methods to compute the symbol error probability in the presence of ISI.
Exhaustive method
We refer to the case of 4-PAM transmission with Ni D 2 interferers due to one non-zero precursor and one non-zero postcursor. Therefore we have ik D ak
1h1
C akC1 h
(7.232)
We dene the vector of symbols that contribute to ISI as a0 D [ak k Then ik can be written as a function of a 0 as k ik D i.a0 / k (7.234)
1 ; akC1 ]
(7.233)
Therefore, ik is a random variable that assumes values in an alphabet with cardinality L D M Ni D 16.
605
Starting from (7.230) the error probability can be computed by conditioning with respect to the values assumed by a0 D [ .1/ ; .2/ ] D 2 A2 . For equally likely symbols and k thresholds given by (7.228) we have h 0 i.a0 / 1 X k P[a0 D ] Pe D 2 1 Q k M 2A2 (7.235) 1 1 X h 0 i./ D2 1 Q M L 2
2A
This method gives the exact value of the error probability in the presence of interferers, but requires the computation of L terms. This method can be costly, especially if the number of interferers is large: it is therefore convenient to consider approximations of the error probability obtained by simpler computational methods.
Gaussian approximation
If interferers have a similar amplitude and their number is large, we can use the central limit theorem and approximate ik as a Gaussian random variable. As the process w R;k is Gaussian, the process z k D ik C w R;k is also Gaussian with variance
2 z2 D i C 2 2 where i is given by (7.72). Then
(7.236)
(7.237) (7.238)
Pe D 2 1
1 M
h0 z
It is seen that this method, although very convenient, is rather pessimistic, especially for large values of 0. As a matter of fact, we observe that the amplitude of ik is limited by the value X imax D .M 1/ jh i j (7.239)
i 6D0
whereas the Gaussian approximation implies that the values of ik are unlimited.
Worst-case bound
This method substitutes ik with the constant imax dened in (7.239). In this case Pe is equal to h 0 imax 1 Q (7.240) Pe D 2 1 M This bound is typically too pessimistic, however, it yields a good approximation if ik is mainly due to one dominant interferer.
606
Saltzberg bound
With reference to (7.230), dening z k as the total disturbance given by (7.236), in general we have 1 P[z k > h 0 ] Pe D 2 1 (7.241) M Let max D maxfn g D M
n
(7.242)
in the specic case, and I be any subset of the integers Z 0 , excluding zero, such that X
i 2I
jh i j <
h0 max
(7.243)
Moreover, let I C be the complementary set of I with respect to Z 0 . Saltzberg applied a Chernoff bound to the probability P[z k > h 0 ] [10], obtaining 0 B B B P[z k > h 0 ] < exp B B B @ !2 1 h 0 max jh i j C C C i 2I 0 1C C C X 2 2 2A A 2 @ C a jh i j X
i 2I C
(7.244)
The bound is particularly simple in the case of binary signaling, where fak g 2 f 1; 1g, 0 B B B Pe < exp B B B @ !2 1 C h0 jh i j C C i 2I 0 1C C C X 2A A @ 2 C 2 jh i j X
i 2I C
(7.245)
P where I is such that i 2I jh i j < h 0 . In this case it is rather simple to choose the set I so that the limit is tighter. We begin with I D Z 0 . Then we remove from I one by one the indices i that correspond to the larger values of jh i j; we stop when the exponent of (7.245) has reached the minimum. Considering the limit of the function Q given by (6.364), we observe that for I D Z 0 and I C D ;, the bound in (7.244) practically coincides with the worst-case limit in (7.240). Taking instead I D ; and I C D Z 0 we obtain again the limit given by the Gaussian approximation for z k that yields (7.238). For the mathematical details we refer to [10]; for a comparison between the Saltzberg bound and other bounds we refer to [5, 11].
607
GQR method
The GQR method is based on a technique for the approximate computation of integrals called Gauss quadrature rule (GQR). It offers a good compromise between computational complexity and approximation accuracy. If we assume a very large number of interferers, to the limit innite, ik can be modelled as a continuous random variable. Then Pe assumes the expression Z C1 1 h0 1 pik . / d D 2 1 I (7.246) Q Pe D 2 1 M M 1 By the GQR method we obtain an approximation of the integral, given by I D
Nw X jD1
wj Q
h0
(7.247)
In this expression the parameters f j g and fw j g are called, respectively, abscissae and weights of the quadrature rule, and are obtained by a numerical algorithm based on the rst 2Nw moments of ik . The quality of the approximation depends on the choice of Nw [5].
608
Appendix 7.C
General scheme
For transmission over a passband channel, a PAM signal must be suitably shifted in frequency by a sinusoidal carrier at frequency f 0 . This task is achieved by DSB modulation (see Example 1.6.3 on page 41) of the signal s.t/ at the output of the baseband PAM modulator lter. In the case of a coherent receiver, the passband scheme is given in Figure 7.52. For the baseband equivalent model, we refer to Figure 7.53a. Now we consider the study of the PAM-DSB transmission system in the unied framework of Figure 7.12. Assuming the receive lter g Rc real-valued, we apply the operator Re [ ] to the channel lter impulse response and to the noise signal, and we split the factor 1=2 evenly among the channel lter and the receive lter responses; setting 1 g Rc .t/ D g Rc .t/ p , we thus obtain the simplied scheme of Figure 7.53b, where the noise 2 signal contains only the in-phase component w0I .t/ with PSD Pw0I . f / D and " gC .t/ D Re or, in the frequency domain, GC . f / D e
j .'1 '0 / G . f Ch
N0 (V2 /Hz) 2 #
(7.248)
p 2 2
(7.249)
C f 0 /1. f C f 0 / C e j .'1 p 4 2
'0 / G . Ch
f C f 0 /1.
f C f0 / (7.250)
For a non-coherent receiver we refer to the scheme developed in Example 6.11.6 on page 516.
609
Consequently the transmitted signal bandwidth is equal to twice the bandwidth of h T x . The minimum bandwidth is given by Bmin D 1 T (7.252)
Recalling the denition (6.103), the spectral efciency of the transmission system is given by D log2 M (bit/s/Hz) which is halved with respect to M-PAM (see Table 6.9). (7.253)
Signal-to-noise ratio
We assume the function e
j .'1 '0 / g .bb/ .t/ Ch
p 2 2
(7.254)
is real-valued; then from Figure 7.53a, using (1.295), we have the following relation: E[jsC .t/j2 ] D
.bb/ E[jsCh .t/j2 ] D E[jsCh .t/j2 ] 2
(7.255)
610
Setting qC .t/ D h T x gC .t/ from (6.105) and (7.252) we have 0D where, for an M-PAM system (6.110), Ma D M2 1 3 (7.258) Ma E qC N0 (7.257) (7.256)
In the absence of ISI, for dened in (7.106), (7.107) still holds; moreover, using (7.257), for a matched lter receiver, (7.113) yields
MF
E qC 20 D N0 =2 Ma r Q ! 1
(7.259)
We observe that the performance of an M-PAM-DSB system and that of an M-PAM system are the same, in terms of Pe as a function of the received power. However, because of DSB modulation, the required bandwidth is doubled with respect to both baseband PAM transmission and PAM-SSB modulation.14 This explains the limited usage of PAM-DSB for digital transmission.
14 The PAM-SSB scheme presents in practice considerable difculties because the lter for modulation is non-
ideal: in fact, this causes distortion of the signal s.t/ at low frequencies that may be compensated for only by resorting to line coding (see Appendix 7.A).
611
Appendix 7.D
Three structures, which differ by the position of the digital-to-analog converter, may be considered for the implementation of a QAM transmitter. In Figure 7.54 the modulator employs for both in-phase and quadrature signals a DAC after the interpolator lter h T x , followed by an analog mixer that shifts the signal to passband. This scheme works if the sampling frequency 1=Tc is much greater than twice the bandwidth B of h T x . For applications where the symbol rate is very high, the DAC is placed right after the bit mapper and the various lters are analog (see Chapter 19). In the implementation illustrated in Figure 7.55, the DAC is placed instead at an intermediate stage with respect to the case of Figure 7.54. Samples are premodulated by a digital mixer to an intermediate frequency f 1 , interpolated by the DAC and subsequently remodulated by a second analog mixer that shifts the signal to the desired band. The intermediate frequency f 1 must be greater than the bandwidth B and smaller than 1=.2Tc / B, thus avoiding overlap among spectral components. We observe that this scheme requires only one DAC, but the sampling frequency must be at least double as compared to the previous scheme.
612
For the rst implementation, as the system is typically oversampled with a sampling interval Tc D T =4 or Tc D T =8, the frequency response of the DAC, G I . f /, may be considered as a constant in the passband of both the in-phase and quadrature signals. For the second implementation, unless f 1 1=Tc , the distortion introduced by the DAC should be considered and equalized by one of these methods (see page 338): including the compensation for G I . f / in the frequency response of the lter h T x , inserting a digital lter before the DAC, inserting an analog lter after the DAC. We recall that an efcient implementation of interpolator lters h T x is obtained by the polyphase representation, as shown in Figure 7.56 for Tc D T =8, where T h .`/ .m/ D h T x mT C ` ` D 0; 1; : : : ; 7 m D 1; : : : ; C1 (7.261) 8 To implement the scheme of Figure 7.56, once the impulse response is known, it may be convenient to precompute the possible values of the lter output and store them in a table or RAM. The symbols fak;I g are then used as pointers for the table itself. The same approach may be followed to generate the values of the signals cos.2 f 1 nTc / and sin.2 f 1 nTc / in Figure 7.55, using an additional table and the index n as a cyclic pointer.
613
Appendix 7.E
In Figure 7.12 we consider the baseband equivalent scheme of a QAM system. The aim is to simulate the various transformations in the discrete-time domain and to estimate the bit error probability. This simulation method, also called Monte Carlo, is simple and general because it does not require any special assumption on the processes involved; however, it is intensive from the computational point of view. For alternative methods, for example semi-analytical, to estimate the error probability, we refer to specic texts on the subject [12]. We describe the various transformations in the overall discrete-time system depicted in Figure 7.57, where the only difference with respect to the scheme of Figure 7.12 is that the
614
lters are discrete-time with quantum TQ D T =Q 0 , Which is chosen to accurately represent the various signals. Binary sequence fb` g. The sequence fb` g is generated as a random sequence or as a PN sequence (see Appendix 3.A), and has length K . Bit mapper. The bit mapper maps patterns of information bits to symbols; the symbol constellation depends on the modulator (see Figure 7.6 for two constellations). Interpolator lter h T x from period T to TQ . The interpolator lter is efciently implemented by using the polyphase representation (see Appendix 1.A). For a bandlimited pulse of the raised cosine or square root raised cosine type, the maximum value of TQ , submultiple of T , is T =2. In any case, the implementation of lters, for example, the lter representing the channel, and non-linear transformations, for example, the transformation due to a power amplier operating near saturation (not considered in Figure 7.57), typically require a larger bandwidth, leading, for example, to the choice TQ D T =4 or T =8. In the following examples we choose TQ D T =4. For the design of h T x the window method can be used (Nh odd): Nh 1 TQ w Nh .q/ q D 0; 1; : : : ; Nh 1 (7.262) h T x .q TQ / D h i d q 2 where typically w Nh is the discrete-time rectangular window or the Hamming window, and h i d is the ideal impulse response. Frequency responses of h T x are illustrated in Figure 7.58 for h i d square root raised cosine pulse with roll-off factor D 0:3, and w Nh rectangular window of length Nh , for various values of Nh (TQ D T =4). The corresponding impulse responses are shown in Figure 7.59. Transmission channel. For a radio channel the discrete-time model of Figure 4.35 can be used, where in the case of channel affected by fading, the coefcients of the FIR lter that model the channel impulse response are random variables with a given power delay prole. For a transmission line the discrete-time model of (4.150) can be adopted. We assume the statistical power of the signal at output of the transmission channel is given by MsCh D MsC . Additive white Gaussian noise. Let w I .q TQ / and w Q .q TQ / be two Gaussian statistically N N independent r.v.s, each with zero mean and variance 1=2, generated according to (1.655). To generate the complex-valued noise signal fwC .q TQ /g with spectrum N0 , it is sufcient to use the relation wC .q TQ / D wC [w I .q TQ / C j w Q .q TQ /] N N (7.263)
615
N = 17 h N = 25 h N = 33
h
10
20
| HT (f) | (dB)
30
x
40
50
60
0.2
0.4
0.6
0.8
1 fT
1.2
1.4
1.6
1.8
Figure 7.58. Magnitude of the transmit lter frequency response, for a windowed square root raised cosine pulse with roll-off factor D 0:3, for three values of Nh (TQ D T=4).
0.3
h (q T )
Nh=17
Q Tx
0.2
0.1
0
0.1
5
q / TQ
10
15
20
0.3
h (q T )
Nh=25
Q Tx
0.2
0.1
0
0.1
10
q / TQ
15
20
25
0.3
h (q T )
Nh=33
Q Tx
0.2
0.1
0
0.1
10
15 q / TQ
20
25
30
Figure 7.59. Transmit lter impulse response, fhTx .qTQ /g, q D 0; : : : ; Nh 1 , for a windowed square root raised cosine pulse with roll-off factor D 0:3, for three values of Nh (TQ D T=4).
616
where
2 wC D N0
1 TQ
(7.264)
Usually the signal-to-noise ratio 0 given by (6.105) is given. For a QAM system, from (7.51) and (7.55) we have MsC MsC D 2 (7.265) 0D N0 .1=T / wC .TQ =T / The standard deviation of the noise to be inserted in (7.263) is given by r MsC Q 0 wC D (7.266) 0 We note that wC is a function of MsC , of the oversampling ratio Q 0 D T =TQ , and of the given ratio 0. In place of 0, the ratio E b =N0 D 0= log2 M may be assigned. Receive lter. As will be discussed in Chapter 8, there are several possible solutions for the receive lter. The most common choice is a matched lter g M , matched to h T x , of the square root raised cosine type. Alternatively, the receive lter may be a simple antialiasing FIR lter g A A , with passband at least equal to that of the desired signal. The lter attenuation in the stopband must be such that the statistical power of the noise evaluated in the passband is larger by a factor of 510 with respect to the power of the noise evaluated in the stopband, so that we can ignore the contribution of the noise in the stopband at the output of the lter g A A . If we adopt as bandwidth of g A A the Nyquist frequency 1=.2T /, the stopband of an ideal lter with unit gain goes from 1=.2T / to 1=.2TQ /: therefore the ripple s in the stopband must satisfy the constraint 1 N0 2T 1 s N0 2T Q from which we get the condition s < 10 Q0
1
1 2T
> 10
(7.267)
(7.268)
Usually the presence of other interfering signals forces the selection of a value of s that is smaller than that obtained in (7.268). Interpolator lter. The interpolator lter is used to increase the sampling rate from 1=TQ 0 to 1=TQ : this is useful when TQ is insufcient to obtain the accuracy needed to represent the timing phase t0 . This lter can be part of g M or g A A . From Appendix 1.A, the efcient 0 0 implementation of fg M . pTQ /g is obtained by the polyphase representation with TQ =TQ branches. To improve the accuracy of the desired timing phase, further interpolation, for example, linear, may be employed.
617
Timing phase. Assuming a training sequence is available, for example, of the PN type fa0 D p.0/; a1 D p.1/; : : : ; a L 1 D p.L 1/g, a simple method to determine t0 is to choose the timing phase in relation to of the peak of the overall impulse response. Let 0 fx. pTQ /g be the signal before downsampling. If we evaluate
0 m opt D arg max jrxa .mTQ /j m
(7.269)
0 0 0 m min TQ < mTQ < m max TQ
(7.270)
In (7.269) and are estimates of minimum and maximum system delay, 0 respectively. Moreover, we note that the accuracy of t0 is equal to TQ and that the amplitude 0 /=r .0/. of the desired signal is h 0 D rxa .m opt TQ a Downsampler. The sampling period after downsampling is usually T or Tc D T =2, with timing phase t0 . The interpolator lter and the downsampler can be jointly implemented, 0 according to the scheme of Figure 1.81. For example, for TQ D T =4, TQ D T =8, and Tc D 0 T =2 the polyphase representation of the interpolator lter with output fx. pTQ /g requires two branches. Also the polyphase representation of the interpolator-decimator requires two branches. Equalizer. After downsampling, the signal is usually input to an equalizer (LE, FSE or DFE, see Chapter 8). The output signal of the equalizer has always sampling period equal to T . As observed several times, to decimate simply means to evaluate the output at the desired instants. Data detection. The simplest method resorts to a threshold detector, with thresholds determined by the constellation and the amplitude of the pulse at the decision point. Viterbi algorithm. An alternative to the threshold detector, which operates on a symbol by symbol basis, is represented by maximum likelihood sequence detection by the Viterbi algorithm (see Chapter 8). Inverse bit mapper. The inverse bit mapper performs the inverse function of the bit mapper. It translates the detected symbols into bits that represent the recovered information bits. Simulations are typically used to estimate the bit error probability of the system, for a certain set of values of 0. We recall that caution must be taken at the beginning and at the N end of a simulation to consider transients of the system. Let K be the number of recovered bits. The estimate of the bit error probability Pbit is given by number of bits received with errors O Pbit D N number of received bits, K (7.271)
618
N O It is known that as K ! 1, the estimate Pbit has a Gaussian probability distribution N N with mean Pbit and variance Pbit .1 Pbit /= K . From this we can deduce, by varying K , O the condence interval [P ; PC ] within which the estimate Pbit approximates Pbit with an assigned probabilty, that is O P[P Pbit PC ] D Pconf (7.272)
N For example, we nd that with Pbit D 10 ` and K D 10`C1 , we have a condence interval O of about a factor 2 with a probability of 95%, that is P[1=2Pbit Pbit 2Pbit ] ' 0:95. This is in good agreement with the experimental rule of selecting N K D 3 10`C1 (7.273)
For a channel affected by fading, the average Pbit is not very signicant: in this case it is meaningful to compute the distribution of Pbit for various channel realizations. In pratice N we assume the transmission of a sequence of N p packets, each one with K p information bits N to be recovered: typically K p D 100010000 bits and N p D 1001000 packets. Moreover, the channel realization changes at every packet. For a given average signal-to-noise ratio O N 0 (see (6.347)), the probability Pbit .n p /, n p D 1; : : : ; N p is computed for each packet. As O a performance measure we use the percentage of packets with Pbit .n p / < Pbit , also called bit error probability cumulative distribution function (cdf), where Pbit assumes values in a certain set. This performance measure is more signicant than the average Pbit evaluated for a very N long, continuous transmission of N p K p information bits. In fact the average Pbit does not show that, in the presence of fading, the system may occasionally have a very large Pbit , and consequently an outage.
Chapter 8
With reference to PAM and QAM systems, in this chapter we will discuss several methods to compensate for linear distortion introduced by the transmission channel. Next, as an alternative to a memoryless threshold detector, we will analyze detection methods that operate on sequences of samples. Recalling the analysis of Section 7.3, we rst review three techniques relying on the zero-forcing lter, linear equalizer, and DFE, respectively, that attempt to reduce the ISI in addition to maximizing the ratio dened in (7.106).
8.1
From the relation (7.66), assuming that HTx . f / and GC . f / are known, and H. f / is given, for example, by (7.84), then the equation H. f / D Q R . f /e j2 f t0 D HTx . f /GC . f /GRc . f /e j2 f t0 can be solved with respect to the receive lter, yielding f ; T rcos 1=T GRc . f / D e HTx . f /GC . f / (8.1)
j2 f t0
(8.2)
From (8.2), the magnitude and phase responses of GRc can be obtained. In practice, however, although the condition (8.2) leads to the suppression of the ISI, hence the lter gRc is called linear equalizer zero-forcing (LE-ZF), it may also lead to the enhancement of the noise power at the decision point, as expressed by (7.75). In fact, if the frequency response GC . f / exhibits strong attenuation at certain frequencies in the range [ .1 C /=.2T /; .1 C /=.2T /], then GRc . f / presents peaks that determine a 2 large value of w R . In any event, the choice (8.2) guarantees the absence of ISI at the decision point, and from (7.109) we get
LE ZF
2 N0 E gRc
(8.3)
620
MF
(8.4)
where MF is dened in (7.113). In the particular case of an ideal channel, that is GCh . f / D G0 in the passband of the system, and assuming h Tx is given by s f (8.5) ; HTx . f / D T rcos 1=T then from (7.42) (8.6) p where from (7.38), k1 D G0 for a PAM system, whereas k1 D .G0 = 2/e j .'1 '0 / for a QAM system. Moreover, from (8.2), neglecting a constant delay, i.e. for t0 D 0, it results that s 1 f (8.7) ; GRc . f / D rcos k1 1=T In other words gRc .t/ is matched to qC .t/ D k1 h Tx .t/, and
LE ZF
QC . f / D HTx . f / GC . f / D k1 HTx . f /
MF
(8.8)
Methods for the design of a LE-ZF lter with a nite number of coefcients are given in Section 8.7.
8.2
We introduce an optimization criterion for GRc that takes into account the ISI as well as the statistical power of the noise at the decision point.
8.2.1
With reference to the scheme of Figure 7.12 for a QAM system, the criterion of choosing the receive lter such that the signal yk is as close as possible to ak in the mean-square sense is widely used.1 Let h Tx and gC be known. Dening the error ek D a k yk yk j2 ] (8.9)
the receive lter gRc is chosen such that the mean-square error J D E[jek j2 ] D E[jak is minimized.
1
(8.10)
It would be desirable to nd the lter such that P[ak 6D ak ] is minimum. This problem, however, is usually O very difcult to solve. Therefore we resort to the criterion of minimizing E[jyk ak j2 ] instead.
621
The following assumptions are made: 1. the sequence fak g is wide sense stationary (WSS) with spectral density Pa . f /; 2. the noise wC is complex-valued and WSS. In particular we assume it is white with spectral density PwC . f / D N0 ; 3. the sequence fak g and the noise wC are statistically independent. The minimization of J in this situation differs from the classical problem of the optimum Wiener lter because h Tx and gC are continuous-time pulses. By resorting to the calculus of variations (see Appendix 8.A), we obtain the general solution GRc . f / D
QC . f / e N0 j2 f t0
Pa . f / 1 T C Pa . f / T
C1 X `D 1
1 QC f N0
` 2 T
(8.11)
where QC . f / D HTx . f /GC . f /. Considerations on the joint optimization of the transmit and receive lters are discussed in Appendix 8.A. 2 If the symbols are statistically independent and have zero mean, then Pa . f / D T a , and (8.11) becomes:
GRc . f / D QC . f /e j2 f t0 2 a C1 1 X 2 QC f N0 C a T `D 1
` 2 T
(8.12)
The expression of the cost function J in correspondence of the optimum lter (8.12) is given in (8.40). From the decomposition (7.62) of GRc . f /, in (8.12) we have the following correspondences:
G M . f / D QC . f /e j2 f t0
(8.13)
and C.e j2 f T / D
2 a C1 1 X 2 QC f N0 C a T `D 1
` 2 T
(8.14)
The optimum receiver thus assumes the structure of Figure 8.1. We note that g M is the matched lter to the impulse response of the QAM system at the receiver input.2 The lter c is called linear equalizer (LE). It attempts to nd the optimum trade-off between removing the ISI and enhancing the noise at the decision point.
As derived later in the text (see Observation 8.13 on page 681) the output signal of the matched lter, sampled at the modulation rate 1=T , forms a sufcient statistic if all the channel parameters are known.
622
Figure 8.1. Optimum receiver structure for a channel with additive white noise.
We analyze two particular cases of the solution (8.12). 1. In the absence of noise, wC .t/ ' 0, and C.e j2 f T / D 1 C1 X 1 QC f T `D 1 ` 2 T (8.15)
Note that the system is perfectly equalized, i.e. there is no ISI. In this case the lter (8.15) is the linear equalizer zero-forcing, as it completely eliminates the ISI. 2. In the absence of ISI at the output of g M , that is if jQC . f /j2 is a Nyquist pulse, then C.e j2 f T / is constant and the equalizer can be removed.
We notice the presence of the parameter D that denotes the lag of the desired signal: this parameter, which must be suitably estimated, expresses in number of symbol intervals the delay introduced by the equalizer. The overall delay from the emission of ak to the generation of the detected symbol ak is equal to t0 C DT seconds. O However, the particular case of a matched lter, for which g M .t/ D qC . .t t0 //, is very interesting from a theoretical point of view. We assume that the lter c may have an innite number of coefcients, i.e. it may be IIR. With reference to the scheme of Figure 8.2a, q is the overall impulse response of the system at the sampler input: q.t/ D h Tx gC g M .t/ D qC g M .t/ D rqC .t t0 / (8.16)
623
(8.17)
The Fourier transform of rqC .t/ is given by PqC . f / D jQC . f /j2 (8.18)
We note that if qC has a nite support .0; tqC /, then g M .t/ D qC .t0 t/ has support .t0 tqC ; t0 /. Hence, to obtain a causal lter g M the minimum value of t0 is tqC . In any case from (8.16), as rqC is a correlation function, the desired sample q.t0 / is taken in relation to the maximum value of jq.t/j. In Figure 8.2a, assuming wC is white noise, we have
(8.19)
(8.20)
624
Then the spectrum of w R is given by: Pw R . f / D N0 PqC . f / D N0 jQC . f /j2 In Figure 8.2a, sampling at instants tk D t0 C kT yields the sampled QAM signal xk D
C1 X iD 1
(8.21)
ai h k
C wk Q
(8.22)
The discrete-time equivalent model is illustrated in Figure 8.2b. The discrete-time overall impulse response is given by h n D q.t0 C nT / D rqC .nT / In particular, it results in h 0 D rqC .0/ D E qC The sequence fh n g has z-transform given by 8.z/ D Z[h n ] D PqC .z/ (8.25) (8.24) (8.23)
which, by the Hermitian symmetry of an autocorrelation sequence, rqC .nT / D rqC . nT /, satises the relation: 1 (8.26) 8.z/ D 8 z
On the other hand, from (1.90), the Fourier transform of (8.23) is given by C1 ` 2 1 X QC f 8.e j2 f T / D F[h n ] D T `D 1 T
(8.27)
Moreover, using the properties of Table 1.3, the correlation sequence of fh n g has z-transform equal to 1 Z[rh .m/] D 8.z/8 (8.28) z Also, from (8.20), the z-transform of the autocorrelation of the noise samples wk D w R .t0 C Q kT / is given by: Z[rw .n/] D Z[rw R .nT /] D N0 8.z/ Q (8.29)
The Wiener solution that gives the optimum coefcients is given in the z-transform domain by (2.50): Copt .z/ D Z[c n ] D where Pdx .z/ D Z[r dx .n/] and Px .z/ D Z[r x .n/] (8.31) Pdx .z/ Px .z/ (8.30)
625
We assume the following assumptions hold: 1. The sequence fak g is WSS, with symbols that are statistically independent and with zero mean,
2 ra .n/ D a n 2 and Pa . f / D T a
(8.32)
2. fak g and fwk g are statistically independent and hence uncorrelated. Q Then the cross-correlation between fdk g and fx k g is given by
rdx .n/ D E[dk x k n ] "
D E ak
C1 X iD 1
C1 X iD 1
! # ai h k
n i
C wk Q
(8.33)
h k
n i E[ak D ai ]
(8.34)
Under the same assumptions 1 and 2, the computation of the autocorrelation of the process fx k g yields (see also Table 1.6):
2 rx .n/ D E[x k x k n ] D a rh .n/ C rw .n/ Q
(8.35)
1 z
z 1 z
2 Px .z/ D a 8.z/8
(8.36) C N0 8.z/
(8.37)
Taking into account the property (8.26), (8.37) is simplied as Copt .z/ D
2 a z D 2 N0 C a 8.z/
(8.38)
626
It can be observed that, for z D e j2 f T , (8.38) corresponds to (8.14), apart from the term z D , which accounts for a possible delay introduced by the equalizer. In relation to the optimum lter C opt .z/, we determine the minimum value of the cost function. We recall the general expression for the Wiener lter (2.53):
2 Jmin D d N 1 X i D0
2 D d
1 2T 1 2T
Z T Z
1 2T 1 2T
1 2T 1 2T
Pdx .e j2 f T /Copt .e j2 f T / d f
(8.40)
2 D a T
N0 df 2 N0 C a 8.e j2 f T /
If 8.z/ is a rational function of z, the integral (8.40) may be computed by evaluating 2 2 the coefcient of the term z 0 of the function a N0 =.N0 C a 8.z//, which can be obtained by series expansion of the integrand, or by using the partial fraction expansion method (see (1.131)). We note that in the absence of ISI, at the output of the MF we get 8.z/ D h 0 D E qC , and Jmin D
2 a N0 2 N0 C a E qC
(8.41)
Signal-to-noise ratio
We dene the overall impulse response at the equalizer output, sampled with a sampling rate equal to the modulation rate 1=T , as
i
D .h n copt;n /i
(8.42)
where fh n g is given by (8.23) and copt;n is the impulse response of the optimum lter (8.38). At the decision point we have yk D
D ak D
C1 X iD 1 i 6D D
i ak D i
C .en copt;n /k w
(8.43)
We assume that in (8.43) the total disturbance given by ISI plus noise is modeled as Gaussian noise with variance 2 I2 . Hence, for a minimum distance among symbols of the constellation equal to 2, (7.106) yields 2 D D (8.44) LE I
627
In case the approximation D ' 1 holds, the total disturbance in (8.43) coincides with ek , hence 2 I2 ' Jmin , and (8.44) becomes
LE
'
2 Jmin
(8.45)
8.3
In practice, if the channel is either unknown a priori or it is time variant, it is necessary to design a receiver that tries to identify the channel characteristics and at the same time to equalize it through suitable adaptive algorithms. Two alternative approaches are usually considered. First solution. The classical block diagram of an adaptive receiver is shown in Figure 8.3. The matched lter g M is designed assuming an ideal channel. Therefore the equalization task is left to the lter c; otherwise, if it is possible to rely on some a priori knowledge of the channel, the lter g M may be designed according to the average characteristics of the channel. The lter c is then an adaptive transversal lter that attempts, in real time, to equalize the channel by adapting its coefcients to the channel variations. Second solution. The receiver is represented in Figure 8.4. The anti-aliasing lter gAA is designed according to specications imposed by the sampling theorem. In particular if the desired signal sC has a bandwidth B and x is sampled with period Tc D T =F0 , where F0 is the oversampling index, with F0 2, then the passband of gAA should extend at least up to frequency B. Moreover, because the noise wC
Figure 8.3. Receiver implementation by an analog matched lter followed by a sampler and a discrete-time linear equalizer.
628
is considered as a wideband signal, g A A should also attenuate the noise components outside the passband of the desired signal sC , hence the cut-off frequency of gAA is between B and F0 =.2T /. In practice, to simplify the implementation of the lter gAA , it is convenient to consider a wide transition band. Thus the discrete-time lter c needs to accomplish the following tasks: 1. to lter the residual noise outside the passband of the desired signal sC ; 2. to act as a matched lter; 3. to equalize the channel. Note that the lter c of Figure 8.4 is implemented as a decimator lter (see Appendix 1.A), where the input signal xn D x.t0 C nTc / is dened over a discrete-time domain with period Tc D T =F0 , and the output signal yk is dened over a discrete-time domain with period T . In turn, two strategies may be used to determine an equalizer lter c with N coefcients: 1. the direct method, which employs the Wiener formulation and requires the computation of the matrix R and the vector p. The description of the direct method is postponed to Section 8.5 (see Observation 8.2 on page 641); 2. the adaptive method, which we will describe next (see Chapter 3).
Adaptive LE
We analyze now the solution illustrated in Figure 8.3: the discrete-time equivalent scheme is illustrated in Figure 8.5, where fh n g is the discrete-time impulse response of the overall system, given by h n D q.t0 C nT / D h Tx gC g M .t/jtDt0 CnT (8.46)
Figure 8.5. Discrete-time equivalent scheme associated with the implementation of Figure 8.3.
629
with w R .t/ D wC g M .t/. The design strategy consists of the following steps. 1. Dene the performance measure of the system. The MSE criterion is typically adopted: J .k/ D E[jek j2 ] (8.48)
2. Select the law of coefcient update. For example, for an FIR lter c with N coefcients using the LMS algorithm (see Section 3.1.2) we have ckC1 D ck C ek x k where a) input vector
T xk D [x k ; x k 1 ; : : : ; x k N C1 ]
(8.49)
(8.50)
b) coefcient vector
T ck D [c0;k ; c1;k ; : : : ; c N 1;k ]
(8.51)
3. To evaluate the signal error ek to be used in the adaptive algorithm we distinguish two modes. a) Training mode ek D a k
D
yk
k D D; : : : ; L TS C D
(8.53)
Evaluation of the error in training mode is possible if a sufciently long sequence of L TS symbols known at the receiver, called training sequence (TS), fak g, k D 0; 1; : : : ; L TS 1, is transmitted. The duration of the transmission of TS is equal to L TS T . During this time interval, the automatic identication of the channel characteristics takes place, allowing the computation of the optimum coefcients of the equalizer lter c, and consequently channel equalization. As the spectrum of the training sequence must be wide, typically a PN sequence is used (see Appendix 3.A). We note that even the direct method requires a training sequence to determine the vector p and the matrix R (see Observation 8.3 on page 641).
630
Figure 8.6. Linear adaptive equalizer implemented as a transversal lter with N coefcients.
yk
k L TS C D
(8.54)
Once the transmission of the TS is completed, we assume that the equalizer has arrived at convergence. Therefore ak ' ak , and the transmission of information symbols may O start. In (8.53) we then substitute the known transmitted symbol with the detected symbol to obtain (8.54). The implementation of the above equations is illustrated in Figure 8.6.
8.4
We consider the receiver structure with oversampling illustrated in Figure 8.4. The discretetime overall system, shown in Figure 8.7, has impulse response given by h i D q.t0 C i Tc / where q.t/ D h Tx gC gAA .t/ The noise is given by wn D w R .t0 C nTc / D wC gAA .t/jtDt0 CnTc Q (8.57) (8.56) (8.55)
631
yn Tc
F0
yk T
^ ak-D T
For the analysis, the lter c is decomposed into a discrete-time lter with sampling period Tc that is cascaded with a downsampler. The input signal to the lter c is the sequence fxn g with sampling period Tc D T =F0 ; the n-th sample of the sequence is given by xn D The output of the lter c is given by
0 yn D N 1 X i D0 C1 X kD 1
hn
k F0 ak
C wn Q
(8.58)
ci xn
(8.59)
We note that the overall impulse response at the lter output, dened on the discrete-time domain with sampling period Tc , is given by
i
D h ci
(8.60)
0 If we denote by fyn g the sequence of samples at the lter output, and by fyk g the downsampled sequence, we have 0 yk D yk F0
(8.61)
0 As mentioned earlier, in a practical implementation of the lter the sequence fyn g is not explicitly generated; only the sequence fyk g is produced (see Appendix 1.A). However, introducing the downsampler helps to illustrate the advantages of operating with an oversampling index F0 > 1. Before analyzing this system, we recall the Nyquist problem. Let us consider a QAM system with pulse h.t/:
s R .t/ D
C1 X nD 1
an h.t
nT /
t 2<
(8.62)
In Section 7.3 we considered continuous-time Nyquist pulses h.t/, t 2 <. Let h.t/ be dened now on a discrete-time domain fnTc g, n integer, where Tc D T =F0 . If F0 is an integer, the discrete-time pulse satises the Nyquist criterion if h.0/ 6D 0, and h.`F0 Tc / D 0, for all integers ` 6D 0. In the particular case F0 D 1 we have Tc D T , and the Nyquist conditions
632
h(t)
H(e j2 fTc )
0 T
2T
t=nTc
(a) F0 D 2.
1 2T
0 1
2T
1 T
h(t)
H(e j2 fTc )
0 T
2T
t=nT
(b) F0 D 1.
1 2T
0 1
2T
1 T
impose that h.0/ 6D 0, and h.nT / D 0, for n 6D 0. Recalling the inputoutput downsampler relations in the frequency domain, it is easy to deduce the behavior of a discrete-time Nyquist pulse in the frequency domain: two examples are given in Figure 8.8, for F0 D 2 and F0 D 1. We note that, for F0 D 1, in the frequency domain a discrete-time Nyquist pulse is equal to a constant. With reference to the scheme of Figure 8.7, the QAM pulse dened on the discrete-time domain with period Tc is given by (8.55), where q.t/ is dened in (8.56). From (8.60), using (8.55), the pulse f i g at the equalizer output before the downsampler has the following Fourier transform: 9.e j2 f Tc / D C.e j2 f Tc /H .e j2 f Tc / C1 X j2 f Tc 1 D C.e / Q f T `D 1 F0 ` T e
F j2 f ` T0 t0
(8.63)
The task of the equalizer is to yield a pulse f i g that approximates a Nyquist pulse, i.e. a pulse of the type shown in Figure 8.8. We see that choosing F0 D 1, i.e. sampling the equalizer input signal with period equal to T , it may happen that H .e j2 f Tc / assumes very small values at frequencies near f D 1=.2T /, because of an incorrect choice of the timing phase t0 . In fact, let us assume q.t/ is real with a bandwidth smaller than 1=T . Using the 1 polar notation for Q. 2T / we have Q 1 2T D Ae j' and Q 1 2T D Ae
j'
(8.64)
633
as Q. f / is Hermitian. Therefore, from (8.63), t0 1 1 e j2 2T C Q H .e j2 f T / DQ 1 f D 2T 2T 2T t0 D 2A cos ' C T If t0 is such that 'C then H .e j2 2T T / D 0
1
1 T
1 j2 2T
1 T t0
(8.65)
2i C 1 t0 D T 2
i integer
(8.66)
(8.67)
In this situation the equalizer will enhance the noise around f D 1=.2T /, or converge with difculty in the adaptive case. If F0 2 is chosen, this problem is avoided because aliasing between replicas of Q. f / does not occur. Therefore the choice of t0 may be less accurate. In fact, as we will see in Chapter 14, if c has an input signal sampled with sampling period T =2 it also acts as an interpolator lter, whose output can be used to determine the optimum timing phase. In conclusion, the FSE receiver presents two advantages over T-spaced equalizers: 1. It is an optimum structure according to the MSE criterion, in the sense that it carries out the task of both matched lter (better rejection of the noise) and equalizer (reduction of ISI). 2. It is less sensible to the choice of t0 . In fact, the correlation method (7.269) with 0 accuracy TQ D T =2 is usually sufcient to determine the timing phase.
Adaptive FSE
The direct method to compute the coefcients of a FSE is described in Section 8.5 (see Observation 8.7 on page 644); we consider now the adaptive method as depicted in Figure 8.9. The choice of the oversampling index F0 D 2 is very common. For this choice, the input samples of the lter c have sampling period T =2, and the output samples have sampling period T . Note that coefcient update takes place every T seconds. With respect to the basic scheme of Figure 8.7, in a practical implementation the equalizer output is not generated at every sampling instant multiple of T =2, but only at alternate sampling instants. The LMS adaptation equation is given by: ckC1 D ck C ek x 2k (8.68)
The adaptive FSE may incur a difculty in the presence of noise with variance that is small with respect to the level of the desired signal: in this case some eigenvalues of the autocorrelation matrix of x may assume a value that is almost zero and consequently 2k
634
x2k T/2
x 2k-1 T/2
x 2k-2 T/2
x 2k-N+1
c 0,k
ACC
c 1,k
ACC
c 2,k
ACC
c N-1,k
ACC
ek
yk
a k-D
Figure 8.9. Adaptive FSE (F0 D 2).
the problem of nding the optimum coefcients become ill-conditioned, with numerous solutions that present the same minimum value of the cost function. This effect can be illustrated also in the frequency domain: outside the passband of the input signal the lter c may assume arbitrary values, in the limit case of absence of noise. As a result, the coefcients of the lter c may vary in time and also assume very large values. To mitigate this problem, recalling the leaky LMS algorithm (see page 187), we consider two methods that slightly modify the cost function. In both cases we attempt to impose a constraint on the amplitude that the coefcients may assume. 1. The leaky LMS algorithm. Let " J1 D J C E then ckC1 D ck C .ek x 2k D .1 with 0 < rx .0/. 2. Let " J2 D J C E
N 1 X i D0 N 1 X i D0
# jci j
2
(8.69)
ck /
/ck C ek x 2k
(8.70)
# jci j (8.71)
635
sgn ck C ek x 2k
For an analysis of the performance of the FSE, including convergence properties, we refer to [1]. Simulations results also demonstrate better performance of the FSE with respect to an equalizer that operates at the symbol rate [2].
8.5
We consider the sampled signal at the output of the analog receive lter (see Figure 8.5 or Figure 8.7). For example, in the scheme of Figure 8.5, the desired signal is given by: sk D s R .t0 C kT / D
C1 X iD 1
ai h k
(8.73)
where the sampled pulse fh n g is dened in (8.46). In the presence of noise we have x k D s k C wk Q (8.74)
where wk is the noise, given by (8.47). Q We assume, as illustrated in Figure 8.10, that fh n g has nite duration and support [ N1 ; N1 C1; : : : ; N2 1; N2 ]. The samples with positive time indices are called postcursors, and those with negative time indices precursors. Explicitly writing terms that include precursors and postcursors, (8.74) becomes: x k D h 0 ak C .h
N1 akCN1
C C h
1 akC1 / C .h 1 ak 1
C C h N 2 ak
Q N 2 / C wk
(8.75)
In addition to the actual symbol ak that we desire to detect from the observation of x k , in (8.75) two terms are identied in parentheses: one that depends only on past symbols ak 1 ; : : : ; ak N2 , and another that depends only on future symbols, akC1 ; : : : ; akCN1 . If the past symbols and the impulse response fh n g were perfectly known, we could use an ISI cancellation scheme limited only to postcursors. Substituting the past symbols with O their detected versions fak 1 ; : : : ; ak N2 g, we obtain a scheme to cancel in part ISI, as O illustrated in Figure 8.11, where, in general, the feedback lter has impulse response fbn g, n D 1; : : : ; M2 , and output given by O xFB;k D b1 ak
1
C C b M2 a k O
M2
(8.76)
If M2 N2 , bn D h n , for n D 1; : : : ; N2 , bn D 0, for n D N2 C 1; : : : ; M2 , and ak i D ak i , for i D 1; : : : ; N2 , then the DFE cancels the ISI due to postcursors. We note O that this is done without changing the noise wk that is present in x k . Q
636
The general structure of a DFE is shown in Figure 8.12, where two lters and the detection delay are outlined: 1. Feedforward (FF) lter c, with M1 coefcients,
M1 1 X i D0
z k D xFF;k D
ci x k
(8.77)
637
Figure 8.11. Simplied scheme of a DFE, where only the feedback lter is included.
bi ak
i D
(8.78)
We recall that for a LE the goal is to obtain a pulse f n g free of ISI, with respect to the desired sample D . Now (see Figure 8.10) ideally the task of the feedforward lter is to obtain an overall impulse response f n D h cn g with very small precursors and a transfer function Z[ n ] that is minimum phase (see Example 1.4.3). In this manner almost all the ISI is cancelled by the FB lter. We note that the FF lter may be implemented as a FSE, whereas the feedback lter operates with sampling period equal to T . The choice of the various parameters depends on fh n g. The following guidelines, however, are usually observed. 1. M1 T =F0 (time-span of the FF lter) at least equal to .N1 C N2 C 1/T =F0 (time-span of h), so the FF lter can effectively equalize;
638
2. M2 T (time-span of the FB lter) equal to or less than .M1 1/T =F0 (time-span of the FF lter minus one); M2 depends also on the delay D, which determines the number of postcursors. 3. For very dispersive channels, for which we have N1 C N2 M1 , it results M2 T M10T . F 4. The choice of DT, equal to the detection delay, is obtained by initially choosing a large delay, D .M1 1/ =F0 , to simplify the FB lter. If the precursors are not negligible, to reduce the constraints on the coefcients of the FF lter the value of D is lowered and the system is iteratively designed. In practice, DT is equal to or smaller than .M1 1/T =F0 .
T For a LE, instead, DT is approximately equal to N 2 1 F0 ; the criterion is that the center of gravity of the coefcients of the lter c is approximately equal to .N 1/=2.
The detection delays discussed above are referred to a pulse fh n g centered in the origin, that does not introduce any delay.
Adaptive DFE
We consider the scheme implemented in Figure 8.13, where the output signal is given by yk D Dening the coefcient vector D [c0 ; c1 ; : : : ; c M1 and the input vector k D [x k ; x k O O O 1 ; : : : ; x k M1 C1 ; ak D 1 ; ak D 2 ; : : : ; ak D M2 ]
T 1 ; b1 ; b2 ; : : : ; b M2 ] T M1 1 X i D0 M2 X jD1
ci x k
b j ak
D j
(8.80)
(8.81)
(8.82)
639
(8.83)
yk
D
(8.84) D ak
D.
We recall that, during the transmission of a training sequence, ak O The LMS adaptation is given by kC1 D k C ek k
(8.85)
yk j2 ]
(8.86)
the Wiener lter theory may be applied to determine the optimum coefcients of the DFE lter in the case ak D ak , with the usual assumptions of symbols that are i.d.d. and O statistically independent of the noise. For a generic sequence fh i g in (8.73), we recall the following results: 1. cross-correlation between ak and x k
2 rax .n/ D a h n
(8.87)
2. autocorrelation of x k
2 rx .n/ D a rh .n/ C rw .n/ Q
(8.88)
h j h j
rw .n/ D N0 rg M .nT / Q
(8.89)
D h cp D
M1 1 X `D0
c` h p
(8.90)
M1 1 X i D0
ci wk Q
M2 X jD1
b j ak
D j
(8.91)
640
Observing (8.91), the optimum choice of the feedback lter coefcients is given by bi D Substitution of (8.92) in (8.80) yields yk D
M1 1 X i D0 i CD
i D 1; : : : ; M2
(8.92)
ci
xk
M2 X jD1
! h jCD i ak
j D
(8.93)
To obtain the WienerHopf solution, the following correlations are needed: " [p] p D E ak
D
xk
M2 X jD1
!# h jCD
p ak D j 2 D a h D p
(8.94) 1
p D 0; 1; : : : ; M1 [R] p;q D E xk
q M2 X j1 D1
h j1 CD
q ak D j1
xk
2 D a N2 X jD N1 p
M2 X j2 D1
h j2 CD ! C rw . p Q q/ 1
p ak D j2
(8.95)
h j h j
. p q/
M2 X jD1
h jCD
q h jCD p
(8.96)
and, from (8.92), the optimum feedback lter coefcients are given by bi D
M1 1 X `D0
copt;` h i CD
i D 1; 2; : : : ; M2
(8.97)
(8.98)
2 D a 1
M1 1 X `D0
641
Observation 8.1 In the particular case in which all the postcursors are cancelled by the feedback lter, that is for M2 C D D N 2 C M1 (8.95) is simplied as
2 [R] p;q D a D X jD N1
(8.99)
hj
qh j p
C rw . p Q
q/
(8.100)
Observation 8.2 The equations to determine copt for a LE are identical to (8.94)(8.98), with M2 D 0. In particular, the vector p in (8.94) is not modied, while the expression of the elements of the matrix R in (8.95) is modied by the terms including the detected symbols. Observation 8.3 For white noise wC , the autocorrelation of wk is proportional to the autocorrelation of the Q receive lter impulse response: consequently, if the statistical power of wk is known, the Q autocorrelation rw .n/ is easily determined. Finally, the coefcients of the channel impulse Q Q response fh n g and the statistical power of wk used in R and p can be determined by the methods given in Appendix 3.B. Observation 8.4 For a LE, the matrix R is Hermitian and Toeplitz, while for a DFE it is only Hermitian; in any case it is (semi-)denite positive. Efcient methods to determine the inverse of the matrix are described in Section 2.3.2. Observation 8.5 The denition of fh n g depends on the value of t0 , which is determined by methods described in Chapter 14. A particularly useful method to determine the impulse response fh n g in wireless systems (see Chapter 18) resorts to a short training sequence to achieve fast synchronization. We recall that a ne estimate of t0;M F is needed if the sampling period of the signal at the MF output is equal to T . The overall discrete-time system impulse response obtained by sampling the output signal of the anti-aliasing lter gAA (see Figure 8.4) is assumed to be known, e.g. by estimation. The sampling period, for example T =8, is in principle determined by the accuracy with which we desire to estimate the timing phase t0;M F at the MF output. To reduce implementation complexity, however, a larger sampling period of the signal at the MF input is considered, for example, T =2. We then implement the MF g M by choosing, among the four polyphase components (see Section 1.A.9 on page 119) of the impulse response, the component with largest energy, thus realizing the MF criterion (see also (8.16)). This is equivalent to selecting among the four possible components with sampling period T =2 of the sampled output signal of the lter gAA the component with largest statistical power. This method is similar to the timing estimator (14.117).
642
The timing phase t0;A A for the signal at the input of g M is determined during the estimation of the channel impulse response. It is usually chosen either as the time at which the rst useful sample of the overall impulse response occurs, or the time at which the peak of the impulse response occurs, shifted by a number of modulation intervals corresponding to a given number of precursors. Note that, if t M F denotes the duration of g M , then the timing phase at the output of g M is given by t0;M F D t0;A A C t M F . The criterion (7.269), according to which t0 is chosen in correspondence of the correlation peak, is a particular case of this procedure. Observation 8.6 In systems where the training sequence is placed at the end of a block of data (see the GSM frame in Appendix 17.A), it is convenient to process the observed signal fx k g starting from the end of the block, lets say from k D K 1 to 0, thus exploiting the knowledge of the training sequence. Now if f n D h cn g and fbn g are the optimum impulse responses if the signal is processed in the forward mode, i.e. for k D 0; 1; : : : ; K 1, it is easy to B B verify that f n g and fbn g, where B is the backward operator dened on page 27, are the optimum impulse responses in the backward mode for k D K 1; : : : ; 1; 0, apart from a constant delay. In fact, if f n g is ideally minimum phase and causal with respect to the B timing phase, now f n g is maximum phase and anticausal with respect to the new instant of optimum sampling. Also the FB lter will be anticausal. In the particular case fh n g is a B correlation sequence, then f n g can be obtained using as FF lter the lter having impulse B response fcn g
ci x2k
M2 X jD1
b j ak
D j
(8.102)
643
Let
p
D h cp D
M1 1 X `D0
c` h p
(8.103)
M1 1 X `D0
c` h 2.i CD/
i D 1; : : : ; M2
(8.104)
ci
x2k
M2 X jD1
! h 2. jCD/ i ak
. jCD/
(8.105)
Using the following relations (see also Example 1.9.10 on page 72) E[ak E[x 2k where, from (8.57), T rw .m/ D N0 rgAA m Q 2 (8.108)
K 2 x2k i ] D a h 2K x2k 2 p ] D a C1 X nD 1 i
(8.106) h 2n
q
h 2n
C rw . p Q
q/
(8.107)
the components of the vector p and the matrix R of the Wiener problem associated with (8.105) are given by
2 [p] p D a h 2D 2 a p
p D 0; 1; : : : ; M1 h 2n q h p 2n
M2 X jD1
1 ! C rw . p Q
(8.109) q/ (8.110)
[R] p;q D
C1 X nD 1
h 2. jCD/ q h jCD/ p 2. 1
p; q D 0; 1; : : : ; M1
644
The feedforward lter is obtained by solving the system of equations R copt D p (8.111)
and the feedback lter is determined from (8.104). The minimum value of the cost function is given by
2 Jmin D a M1 1 X `D0
(8.112)
2 D a 1
M1 1 X `D0
A problem encountered with this method is the inversion of the matrix R in (8.111), because it may be ill-conditioned. Similarly to the procedure outlined on page 187, a solution consists in adding a positive constant to the elements on the diagonal of R, so that R becomes invertible; obviously the value of this constant must be rather small, so that the performance of the optimum solution does not change signicantly. Observation 8.7 Observations similar to the observations 8.18.3 hold for a FS-DFE, with appropriate changes. In this case the timing phase t0 after the lter gAA can be determined with accuracy T =2, for example by the correlation method (7.269). For an FSE, or FS-LE, the equations to determine copt are given by (8.109)(8.111) with M2 D 0. Note that the matrix R is Hermitian but in general it is no longer Toeplitz. Observation 8.8 Two matrix formulations of the direct method to determine the coefcients of a DFE and an FS-DFE are given in Appendix 8.B. In particular, a formulation uses the correlation of the equalizer input signal fxn g, the correlation of the sequence fak g, and the cross-correlation of the two signals: using suitable estimates of the various correlations (see the correlation method and the covariance method considered in Section 2.3), this method avoids the need for the estimate of the overall channel impulse response; however, it requires a greater computational complexity with respect to the method described in this section.
Signal-to-noise ratio
Using FF and FB lters with an innite number of coefcients, it is possible to achieve the minimum value of Jmin . Salz derived the expression of Jmin for this case, given by [3] 1 0 Z 1 2T N0 2 Jmin D a exp @T dfA (8.113) ln 2 1 N0 C a 8.e j2 f T /
2T
645
e f .a/ da
(8.114)
to (8.113), we can compare the performance of a linear equalizer given by (8.40) with that of a DFE given by (8.113): the result is that, assuming 8.e j2 f T / 6D constant and the absence of detection errors in the DFE, for innite-order lters the value Jmin of a DFE is always smaller than Jmin of a LE. If FF and FB lters with a nite number of coefcients are employed, in analogy with (8.44), also for a DFE we have 2
DFE
Jmin
(8.115)
where Jmin is given by (8.98). An analogous relation holds for an FS-DFE, with Jmin given by (8.112).
Remarks
1. In the absence of errors of the data detector, the DFE has better asymptotic (M1 ; M2 ! 1) performance than the linear equalizer. However, for a given nite number of coefcients M1 C M2 , the performance is a function of the channel and of the choice of M1 . 2. The DFE is denitely superior to the linear equalizer for channels that exhibit large variations of the attenuation in the passband, as in such cases the linear equalizer tends to enhance the noise. 3. Detection errors tend to spread, because they produce incorrect cancellations. Error propagation leads to an increase of the error probability. However, simulations indicate that for typical channels and symbol error probability smaller than 5 10 2 , error propagation is not catastrophic. 4. For channels with impulse response fh i g, such that detection errors may spread catastrophically, instead of the DFE structure it is better to implement the linear FF equalizer at the receiver, and the FB lter at the transmitter as a precoder, using the precoding method discussed in Appendix 7.A and Chapter 13.
8.6
We consider the digital transmission model of Figure 8.5, and observe the performance of adaptive LE and DFE by resorting to a specic channel realization. In particular, we analyze two cases in which the discrete-time overall impulse response of the system, fh n g, is either minimum phase, h min , as in Figure 1.15a, or non-minimum phase, h nom , as in 2 Figure 1.15c. The additive channel noise wk is AWGN with statistical power w , such that Q Q
646
2 2 the signal-to-noise ratio 0 D a rh .0/=w at the equalizer input is equal to 20 dB. The Q sequence of symbols ak 2 f 1; 1g is a PN training sequence of length L D 63.
Adaptive LE
With reference to the scheme of Figure 8.6, we consider a LE with N D 15 coefcients. In terms of mean-square error at convergence, the best results are obtained for a delay D D 0 in the case of h min and D D 8 in the case of h nom ; we observe that the overall impulse response h nom is not centered at the origin, and the delay D is the sum of the delays introduced by fh n g and fcn g. Figures 8.15c, d and 8.16c, d show curves of mean-square error convergence for standard LMS and RLS algorithms (see Section 3.1.2) for minimum and non-minimum phase channels, respectively; in the plots, Jmin represents the minimum value of J achieved with optimum coefcients computed by the direct method. We note that Jmin is 4 dB higher 2 than the value given by w , because of the noise and residual ISI at the decision point. Q The impulse response fcopt;n g of the optimum LE and the overall system impulse response f n D h copt;n g are depicted in Figures 8.15a, b and 8.16a, b for the two channels. The curves of convergence of J .k/ indicate that the RLS algorithm succeeds in achieving convergence by the end of the training sequence, whereas the LMS algorithm still presents a considerable offset from the optimum conditions, even though a large adaptation gain is chosen.
(a)
1.5
1
(b)
1.5
1
|c
0.5
0
|n|
15
opt,n
0.5
0
10
10
15
n
0
(c)
n LMS
J(k) (dB)
5
10
15
20
0
min
10
20
30
40
50
60
(d)
15
J(k) (dB)
20
25
30
35
0
Jmin RLS
10
20
30
40
50
60
Figure 8.15. System impulse responses and curves of mean-square error convergence, estimated over 500 realizations, for a channel with minimum phase impulse response, and a LE employing the LMS with D 0:062 or the RLS.
647
(a)
2
1.5
(b)
2
1.5
| copt,n |
|n|
0.5
0
0
0.5
5
10
15
10
15
n
0
(c)
J(k) (dB)
LMS
5
10
15
20
0
Jmin
10
20
30
40
50
60
(d)
15
Jmin RLS
J(k) (dB)
20
25
30
35
0
10
20
30
40
50
60
Figure 8.16. System impulse responses and curves of mean-square error convergence, estimated over 500 realizations, for a channel with non-minimum phase impulse response, and a LE employing the LMS with D 0:343 or the RLS.
(a)
1 1
(b)
|c
0.5
|n|
10
opt,n
0.5
n
0
(c)
10
n
LMS
J(k) (dB)
5
10
15
20
0
Jmin
10
20
30
40
50
60
(d)
15
J(k) (dB)
Jmin
20
25
30
35
0
RLS
10
20
30
40
50
60
Figure 8.17. System impulse responses and curves of mean-square error convergence, estimated over 500 realizations, for a channel with minimum phase impulse response, and a DFE employing the LMS with D 0:063 or the RLS.
648
(a)
2
(b)
2
1.5
| copt,n |
1.5
1
|n|
0.5
0
0
0.5
5
10
n
0
(c)
10
n
LMS
J(k) (dB)
5
10
15
20
0
Jmin
10
20
30
(d)
40
50
60
15
Jmin
J(k) (dB)
20
25
30
35
0
RLS
10
20
30
40
50
60
Figure 8.18. System impulse responses and curves of mean-square error convergence, estimated over 500 realizations, for a channel with non-minimum phase impulse response, and a DFE employing the LMS with D 0:143 or the RLS.
Adaptive DFE
We consider now the performance of a DFE as illustrated in Figure 8.13, with parameters M1 D 10, M2 D 5, and D D 9, for both h min and h nom . Also in this case the chosen value of D gives the best results in terms of the value of J at convergence. Figures 8.17c, d and 8.18c, d show curves of mean-square error convergence for standard LMS and RLS algorithms, for minimum and non-minimum phase channels, respectively; The impulse response fcopt;n g of the optimum FF lter and the overall system impulse response f n D h copt;n g are depicted in Figures 8.17a, b and 8.18a, b, for the two channels.
8.7
Ignoring the noise, the signal at the output of a LE with N coefcients (see (8.91)) is given by
N 1 X i D0 CN N2X 1 pD N1
yk D
ci x k
p ak p
(8.116)
649
c` h p
D c0 h p C c1 h p pD
C C cN
1 h p .N 1/
(8.117)
N1 ; : : : ; 0; : : : ; N 2 C N
D p
(8.118)
where D is a suitable delay. If the overall impulse response fh n g, n D N1 ; : : : ; N2 , is known, a method to determine the coefcients of the LE-ZF consists in considering the system (8.117) with Nt D N1 C N2 C N equations and N unknowns, that can be solved by the method of the pseudoinverse (see (2.185)); alternatively, the solution can be found in the frequency domain by taking the Nt -point DFT of the various signals (see (1.108)), and the result windowed in the time domain so that the lter coefcients are given by the N consecutive coefcients that maximize the energy of the lter impulse response. An approximate solution is obtained by forcing the condition (8.118) only for N values of p in (8.117) centered around D; then the matrix of the system (8.117) is square and, if the determinant is different from zero, it can be inverted. Note that all these methods require an accurate estimate of the overall impulse response, otherwise the equalizer coefcients may deviate considerably from the desired values. An alternative robust method, which does not require the knowledge of fh n g and can be extended to FSE-ZF systems, will be presented in Section 8.15. An adaptive ZF equalization method is discussed in Appendix 8.C.
8.8
We determine the expressions of the FF and FB lters of a DFE in the case of IIR lter structure.
DFE-ZF
We consider a receiver with a matched lter g M (see Figure 8.1) followed by the DFE illustrated in Figure 8.19, where, to simplify the notation, we assume t0 D 0 and D D 0. The z-transform of the QAM system impulse response is given by 8.z/, as dened in (8.25). With reference to Figure 8.19, the matched lter output x k is input to a linear equalizer zero forcing (LE-ZF) with transfer function 1=8.z/ to remove the ISI: therefore the LE-ZF output is given by x E;k D ak C w E;k (8.119)
650
From (8.29), using the property (8.26), we obtain that the spectrum Pw E .z/ of w E;k is given by Pw E .z/ D N 0 8.z/ 1 8.z/8 1 z (8.120)
D N0
1 8.z/
As 8.z/ is the z-transform of a correlation sequence, it can be factorized as (see page 53) 1 (8.121) 8.z/ D F.z/ F z where F.z/ D
C1 X nD0
fn z
(8.122)
is a minimum-phase function, that is with poles and zeros inside the unit circle, associated with a causal sequence ffn g. Observation 8.9 A useful method to determine the lter F.z/ in (8.121), with a computational complexity that is proportional to the square of the number of lter coefcients, is obtained by considering a minimum-phase prediction error lter A.z/ D 1 C a 0 z 1 C C a0N ;N z N (see 1;N page 147), designed using the ACS frqC .nT /g dened by (8.23). The equation to determine the coefcients of A.z/ is given by (2.85), where fr x .n/g is now substituted by frqC .nT /g. The nal result is F.z/ D f 0 =A.z/. On the other hand, F .1=z / is a function with zeros and poles outside the unit circle, associated with an anticausal sequence f f n g.
651
We choose as transfer function of the lter w in Figure 8.19 the function W .z/ D F.z/ 1 f0 (8.123)
The ISI term in z k is determined by W .z/ 1, hence there are no precursors; the noise is white with statistical power.N0 =f2 /. Therefore the lter w is called whitening lter (WF). 0 In any case, the lter composed of the cascade of LE-ZF and w is also a WF. If ak D ak then, for O B.z/ D 1 W .z/ (8.124)
the FB lter removes the ISI present in z k and leaves the white noise unchanged. As yk is not affected by ISI, this structure is called DFE-ZF, for which we obtain D 2jf0 j2 N0 (8.125)
DFE ZF
With this lter the noise in z k is white. The relation between ak and the desired signal in z k is instead governed by 9.z/ D F.z/=f 0 and B.z/ D 1 9.z/. In other words, the overall discrete-time system is causal and minimum phase, that is the energy of the impulse response is mostly concentrated at the beginning of the pulse. The overall receiver structure is illustrated in Figure 8.20, where the block including the matched lter, sampler, and whitening lter, is called whitened matched lter (WMF). Note that the impulse response at the WF output has no precursors. In principle, the WF of Figure 8.20 is non-realizable, because it is anticausal. In practice we can implement it in two ways:
652
a) by introducing an appropriate delay in the impulse response of an FIR WF, and processing the output samples in the forward mode for k D 0; 1; : : : ; b) by processing the output samples of the IIR WF in the backward mode, for k D K 1; K 2; : : : ; 0, starting from the end of the block of samples. We observe that the choice F.z/ D f 0 =A.z/, where A.z/ is discussed in Observation 8.9 on page 650, leads to an FIR WF with transfer function 12 A .1=z /.
f0
Observation 8.10 With reference to the scheme of Figure 8.20, using a LE-ZF instead of a DFE structure means that a lter with transfer function f0 =F.z/ is placed after the WF to produce the signal x E;k given by (8.119). For a data detector based on x E;k , the ratio is
LE ZF
2 rw E .0/
(8.127)
where rw E .0/ is determined as the coefcient of z 0 in N0 =8.z/. This expression is alternative to (8.3). Example 8.8.1 (WF for a channel with exponential impulse response) A method to determine the WF in the scheme of Figure 8.20 is illustrated by an example. Let p (8.128) qC .t/ D E qC 2e t 1.t/ be the overall system impulse response at the MF input; in (8.128) E qC is the energy of qC . The autocorrelation of qC , sampled at instant nT, is given by rqC .nT / D E qC a jnj Then 8.z/ D Z[r qC .nT /] D E qC D E qC .1 a 2 / az 1 C .1 C a 2 / .1 .1 a 2 / az 1 /.1 az/ az (8.130) aDe
T
<1
(8.129)
We note that the frequency response of (8.129) is 8.e j2 f T / D E qC and presents a minimum for f D 1=.2T /. .1 a 2 / 1 C a 2 2a cos.2 f T / (8.131)
653
With reference to the factorization (8.121), it is easy to identify the poles and zeros of 8.z/ inside the unit circle, hence F.z/ D D q q E qC .1 E qC .1 a2/ a2/ 1 az a z
n
1
C1 X nD0
(8.132)
n
In particular, the coefcient of z 0 is given by f0 D The WF of Figure 8.20 is expressed as 1 f0 F D 1 .1 E qC .1 a 2 / az/ (8.134) 1 z. a C z D E qC .1 a 2 /
1
E qC .1
a2/
(8.133)
1 z
In this case the WF, apart from a delay of one sample .D D 1/, can be implemented by a simple FIR with two coefcients, whose values are equal to a=.E qC .1 a 2 // and 1=.E qC .1 a 2 //. The FB lter is a rst-order IIR lter with transfer function 1 1 F.z/ D f0 D1 D
C1 X nD1
an z 1 az
(8.135)
1
az 1 1 az 1
Example 8.8.2 (WF for a two-ray channel) In this case we directly specify the autocorrelation sequence at the matched lter output: 1 (8.136) 8.z/ D Q CC .z/Q CC z where Q CC .z/ D p E qC .q0 C q1 z
1
(8.137)
654
with q0 and q1 such that jq0 j2 C jq1 j2 D 1 jq0 j > jq1 j (8.138) (8.139)
In this way E qC is the energy of fqCC .nT /g and Q CC .z/ is minimum phase. Equation (8.137) represents the discrete-time model of a wireless system with a two-ray channel. The impulse response is given by p (8.140) qCC .nT / D E qC .q0 n C q1 n 1 / The frequency response is given by QCC . f / D p E qC .q0 C q1 e
j2 f T
(8.141)
We note that if q0 D q1 , the frequency response has a zero for f D 1=.2T /. From (8.136) and (8.137) we get 8.z/ D E qC .q0 C q1 z hence, recalling assumption (8.139), p F.z/ D E qC .q0 C q1 z and f0 D The WF is given by 1 f0 F where bD q1 q0 D 1 E qC jq0 j2 .1 (8.145) p E qC q 0 (8.144)
1 /.q0 C q1 z/
(8.142)
/ D Q CC .z/
(8.143)
1 z
bz/
We note that the WF has a pole for z D b 1 , which, recalling (8.139), lies outside the unit circle. In this case, in order to have a stable lter, it is convenient to associate the z-transform 1=.1 bz/ with an anticausal sequence, 1 1 bz D
0 X iD 1
.bz/
1 X nD0
.bz/ n
(8.146)
655
On the other hand, as jbj < 1, we can approximate the series by considering only the rst .N 1/ terms, obtaining 1 f0 F '
N X 1 .bz/ n E qC jq0 j2 nD0
1 z
(8.147) D 1 z N [b N C C bz E qC jq0 j2
.N 1/
Cz
Consequently the WF, apart from a delay D D N , can be implemented by an FIR lter with N C 1 coefcients. The FB lter in this case is a simple FIR lter with one coefcient 1 1 F.z/ D f0 q1 z q0
1
(8.148)
W .z//.A.z/
the scheme of Figure 8.19 is redrawn as in Figure 8.21, where the FB lter acts as a noise predictor. In fact, for ak D ak , the FB lter input is colored noise. By removing the O correlated noise from x E;k , we obtain yk that is composed of white noise, with minimum variance, plus the desired symbol ak .
Figure 8.21. Predictive DFE: the FF lter is a linear equalizer zero forcing.
656
Figure 8.22. Predictive DFE with the FF lter as a minimum-MSE linear equalizer.
by (8.38). The z-transform of the overall impulse response at the FF lter output is given by: 8.z/C.z/ D As t0 D 0, the ISI in z k is given by: 8.z/C.z/ 1D N0 2 N0 C a 8.z/ N0
2 N0 C a 8 2 a 8.z/ 2 N0 C a 8.z/
(8.150)
(8.151)
Hence, the spectral density of the ISI has the following expression: PI S I .z/ D Pa .z/ N0 2 N0 C a 8.z/ 1 z (8.152)
2 using (8.26) and the fact that the symbols are uncorrelated with Pa .z/ D a . The spectrum of the noise in z k is given by 1 Pnoise .z/ D N 0 8.z/C.z/C z 4 8.z/ a D N0 2 .N0 C a 8.z// 2
(8.153)
Therefore the spectrum of the disturbance vk in z k , composed of ISI and noise, is given by Pv .z/ D
2 N0 a 2 N0 C a 8.z/
(8.154)
We note that the FF lter could be an FSE and the result (8.154) would not change.
657
To minimize the power of the disturbance in yk , the FB lter, with input ak z k D O ak z k D vk (assuming ak D ak ), needs to remove the predictable components of z k . For O a predictor of innite length, we set B.z/ D An alternative form is B.z/ D 1 where A.z/ D
C1 X nD0 0 an z n C1 X nD1
bn z
(8.155)
A.z/
(8.156)
a0 D 1 0
(8.157)
is the forward prediction error lter dened in (2.83). To determine B.z/ we use the spectral factorization in (1.526): Pv .z/ D
2 y 1 A.z/A z 2 y
(8.158) B 1 z
D [1
B.z/] 1
with Pv .z/ given by (8.154). In conclusion, it results that the prediction error signal yk is a white noise process with 2 statistical power equal to y . An adaptive version of the basic scheme of Figure 8.22 suggests that the two lters, c and b, are separately adapted through the error signals fe F;k g and fe B;k g, respectively. This conguration, although sub-optimum with respect to the DFE, is used in conjunction with trellis-coded modulation (see Chapter 12) [4].
8.9
We compare limits on the performance of the two equalizers, LE-ZF and DFE-ZF, in terms of the signal-to-noise ratio at the decision point, .
Performance comparison
From (8.120), the noise sequence fw E;k g can be modeled as the output of a lter having transfer function C F .z/ D 1 F.z/ (8.159)
658
and input given by white noise with PSD N0 . Because F.z/ is causal, also C F .z/ is causal: C F .z/ D
1 X nD0
c F;n z
(8.160)
where fc F;n g, n 0, is the lter impulse response. Then we can express the statistical power of w E;k as: rw E .0/ D N0 where, from (8.159), c F;0 D C.1/ D Using the inequality
1 X nD0 1 X nD0
jc F;n j2
(8.161)
1 1 D F.1/ f0 1 jf0 j2
(8.162)
jc F;n j2 jc F;0 j2 D
(8.163)
DFE ZF
(8.164)
N0 1 C a 2 E qC 1 a 2 E qC 1 a 2 N0 1 C a 2
(8.165)
D2
(8.166)
MF
LE ZF
MF
(8.167)
Therefore the loss due to the ISI, given by the factor .1 a 2 /=.1 C a 2 /, can be very large if a is close to 1. In this case the frequency response in (8.131) assumes a minimum value close to zero.
659
Two-ray channel. From (8.142) we have 1 1 D 8.z/ E qC .q0 C q1 z 1 /.q0 C q1 z/ By a partial fraction expansion, we nd only the pole for z D circle, hence rw E .0/ D (8.168) q 1 =q0 lies inside the unit
N0 N0 D q1 E qC .jq0 j2 jq1 j2 / E qC q 0 q 0 q 1 q0
(8.169)
Then
LE ZF
2 MF .jq0 j
jq1 j2 /
(8.170)
Also in this case we nd that the LE is unable to equalize channels with a spectral zero. DFE-ZF Channel with exponential impulse response. Substituting the expression of f0 given by (8.133) in (8.125), we get
DFE Z F
D D
2 E q .1 N0 C
MF .1 2
a2/ (8.171)
a /
We note that
DFE Z F
is better than
LE ZF
by the factor .1 C a 2 /.
2 E q jq0 j2 D N0 C
2 MF jq0 j
(8.172) jq1 j2 /,
In this case the advantage with respect to LE-ZF is given by the factor jq0 j2 =.jq0 j2 which may be substantial if jq1 j ' jq0 j.
We recall that in case E qC N0 , that is for low noise levels, the performance of LE and DFE are similar to the performance of LE-ZF and DFE-ZF, respectively. Anyway, for the two systems of Examples 8.8.1 and 8.8.2 the values of in terms of Jmin are given in [4], for both LE and DFE.
8.10
Adopting an MSE criterion at the decision point, we have derived the conguration of Figure 8.1 for an LE, and that of Figure 8.12 for a DFE. In both cases, the decision on a transmitted symbol ak D is based only on yk through a memoryless threshold detector.
660
Actually, the decision criterion that minimizes the probability that an error occurs in the detection of a symbol of the sequence fak g requires in general that the entire sequence of received samples is considered for symbol detection. We assume a sampled signal having the following structure: z k D u k C wk where: u k is the desired signal that carries the information, uk D
L2 X nD L 1
k D 0; 1; : : : ; K
(8.173)
n ak
(8.174)
where 0 is the sample of the overall system impulse response, obtained in correspondence of the optimum timing phase; f L 1 ; : : : ; 1 g are the precursors, which are typically negligible with respect to 0 . Recalling the expression of the pulse f p g given by (8.90), we have n D nCD . We assume the coefcients fn g are known; in practice, however, they are estimated by the methods discussed in Appendix 3.B; wk is a circularly symmetric white Gaussian noise, with equal statistical power in 2 the two I and Q components given by I2 D w =2; hence the samples fwk g are uncorrelated and therefore statistically independent. In this section, a general derivation of optimum detection methods is considered; possible applications span, e.g., decoding of convolutional codes (see Chapter 11), coherent demodulation of CPM signals (see Appendix 18.A), and obviously detection of sequences transmitted over channels with ISI. We introduce the following vectors with K components (K may be very large): 1. Sequence of transmitted symbols, or information message, modeled as a sequence of r.v.s from a nite alphabet a D [a L 1 ; a L 1 C1 ; : : : ; a L 1 CK
1] T
ai 2 A
(8.175)
2. Sequence of detected symbols, modeled as a sequence of r.v.s from a nite alphabet O a D [a L 1 ; a L 1 C1 ; : : : ; a L 1 CK O O O 3. Sequence of detected symbol values D [ L 1 ; L 1 C1 ; : : : ; L 1 CK
1] T 1] T
ai 2 A O
(8.176)
i 2 A
(8.177)
(8.178)
661
i 2 C
2 CK
(8.179)
Let M be the cardinality of the alphabet A. By analogy with the analysis of Section 6.1, we divide the vector space of the received samples, C K , into M K non-overlapping regions R 2 AK (8.180)
such that, if belongs to R , then the sequence is detected: O if 2 R H) a D The probability of a correct decision is computed as follows: P[C] D P[O D a] a X D P[O D j a D ]P[a D ] a
2A K
(8.181)
(8.182)
2A K
X Z
R
2A K
As in (6.18), the following criteria may be adopted to minimize (8.182). Maximum a posteriori probability (MAP) criterion 2 R Using the identity pzja . j / D pz ./ the decision criterion becomes: O a D arg max P[a D j z D ]
(8.183)
P[a D j z D ] P[a D ]
(8.184)
(8.185)
An efcient realization of the MAP criterion will be developed in Section 8.10.2. If all data sequences are equally likely, the MAP criterion coincides with the maximum likelihood sequence detection (MLSD) criterion O a D arg max pzja . j /
(8.186)
In other words, the sequence is chosen, for which the probability to observe z D is maximum.
662
8.10.1
We now discuss a computationally efcient method to nd the solution indicated by (8.186). As the vector a of transmitted symbols is hypothesized to assume the value , both a and u D [u 0 ; : : : ; u K 1 ]T are xed. Recalling that the noise samples are statistically independent, from (8.173) we get:3 pzja . j / D
K 1 Y kD0 2 As wk is a complex-valued Gaussian r.v. with zero mean and variance w , it follows
pzk ja .k j /
(8.187)
pzja . j / D
K 1 Y kD0
1 e 2 w
1 u j2 2 j w k k
(8.188)
Taking the logarithm, which is a monotonic increasing function, of both members we get ln pzja . j / /
K 1 X kD0
jk
u k j2
(8.189)
where non-essential constant terms have been neglected. Then the MLSD criterion is formulated as O a D arg min
K 1 X kD0
jk
u k j2
(8.190)
where u k , dened by (8.174), is a function of the transmitted symbols expressed by the general relation u k D fQ.akCL 1 ; : : : ; ak ; : : : ; ak
L2 /
(8.191)
We note that (8.190) is a particular case of the minimum distance criterion (6.30), and it suggests a detecting the vector u that is closest to the observed vector . However, we are interested in detecting the symbols fak g and not the components fu k g. A direct computation method requires that, given the sequence of observed samples, for each possible data sequence of length K , the corresponding K output samples, elements of the vector u, should be determined, and the relative distance, or metric, should be computed as 0.K 1/ D
K 1 X kD0
jk
u k j2
(8.192)
The detected sequence is the sequence that yields the smallest value of 0.K 1/; as in the case of i.i.d. symbols there are M K possible sequences, this method has a complexity O.M K /.
3
We note that (8.187) formally requires that the vector a is extended to include the symbols a L 2 ; : : : ; a L 1 1 .
663
for each possible pair of distinct and in A K . As the noise is white, we dene
2 dmin D min d 2 .u./; u.// ;
(8.194)
Then the lower limit (6.86) can be used, and we get Nmin dmin Pe Q MK 2 I
(8.195)
where M K is the number of vectors u.a/, and Nmin is the number of vectors u.a/ whose distance from another vector is dmin . In practice, the exhaustive method for the computation of the expressions in (8.192) and (8.194) is not used; the Viterbi algorithm, that will be discussed in the next section, is utilized instead.
(8.196)
and the output is given in general by (8.191). We denote by S the set of the states, that is the set of possible values of sk : sk 2 S D f 1 ; 2 ; : : : ; Ns g (8.197)
With the assumption of i.i.d. symbols, the number of states is equal to Ns D M L 1 CL 2 . Observation 8.11 We denote by s 0 k ak L 2 . Then
(8.198)
1
as (8.199)
664
Dening the metric 0k D the following recursive equation holds: 0k D 0k or, using (8.199), 0k D 0k
1 1 k X i D0
ji
u i j2
(8.200)
C jk
u k j2
(8.201)
C jk
f .sk ; sk
2 1 /j
(8.202)
Thus we have interpreted fu k g as the output sequence of a nite state machine, and we have expressed recursively the metric (8.192). Note that the metric is a function of the sequence of states s0 ; s1 ; : : : ; sk , associated with the sequence of output samples u 0 ; u 1 ; : : : ; u k . The following example illustrates how to describe the transitions between states of the nite state machine. Example 8.10.1 Let us consider a transmission system with symbols taken from a binary alphabet, that is M D 2, ak 2 f 1; 1g, and overall impulse response characterized by L 1 D L 2 D 1. In this case we have sk D .akC1 ; ak / and the set of states contains Ns D 22 D 4 elements: S D f 1 D . 1; 1/; 2 D . 1; 1/; 3 D .1; 1/; 4 D .1; 1/g The possible transitions sk
1
(8.203)
(8.204)
D i
! sk D j
(8.205)
are represented in Figure 8.23, where a dot indicates a possible value of the state at a certain instant, and a branch indicates a possible transition between two states at consecutive instants. According to (8.198), the variable that determines a transition is akCL 1 . Figure 8.23, extended for all instants k, is called a trellis diagram. We note that in this case there are exactly M transitions that leave each state sk 1 ; likewise there are M transitions that arrive to each state sk . With each state j , j D 1; : : : ; Ns , at instant k we associate two quantities: 1. the path metric, or cost function, dened as: 0.sk D j / D
s0 ;s1 ;:::;sk D j
min
0k
(8.206)
2. the survivor sequence, dened as the sequence of symbols that ends in that state and determines 0.sk D j /: L.sk D j / D .s0 ; s1 ; : : : ; sk D j / D .a L 1 ; : : : ; akCL 1 / (8.207)
665
Figure 8.23. Portion of the trellis diagram showing the possible transitions from state sk 1 to state sk , as a function of the symbol akCL1 2 A.
Note that the notion of survivor sequence can be equivalently applied to a sequence of symbols or to a sequence of states. These two quantities are determined recursively. In fact, it is easy to verify that if, at instant k, a survivor sequence of states includes sk D j then, at instant k 1, the same sequence includes sk 1 D iopt , which is determined as follows: iopt D arg The term jk
sk
1 D i 2S !sk D j
min
0.sk
D i / C jk
f . j ; i /j2
(8.208)
D iopt / C jk
f . j ; iopt /j2
(8.209)
D iopt /; j /
(8.210)
Starting from k D 0, with initial state s 1 , which may be known or arbitrary, the procedure is repeated until k D K 1. The optimum sequence of states is given by the survivor sequence L.s K 1 D jopt / associated with s K 1 D jopt having minimum cost. If the state s 1 is known and equal to i0 , it is convenient to assign to the states s 1 the following costs: ( 0 for i D i0 0.s 1 D i / D (8.211) 1 otherwise Analogously, if the nal state s K is equal to f 0 , the optimum sequence of states coincides with the survivor sequence associated with the state s K 1 D j having minimum cost among those that admit a transition into s K D f 0 .
666
Example 8.10.2 Let us consider a system with the following characteristics: ak 2 f 1; 1g, L 1 D 0, L 2 D 2, K D 4, and s 1 D . 1; 1/. The development of the survivor sequences on the trellis diagram from k D 0 to k D K 1 D 3 is represented in Figure 8.24a. The branch metric, f . j ; i /j2 , associated with each transition is given in this example and is written jk above each branch. The survivor paths associated with each state are represented in bold; we note that some paths are abruptly interrupted and not extended at the following instant: for example the path ending at state 3 at instant k D 1. Figure 8.24b illustrates how the survivor sequences of Figure 8.24a are determined. Starting with s 1 D . 1; 1/ we have 0.s 0.s 0.s 0.s
1 1 1 1
D 1/ D 0 D 2/ D 1 D 3/ D 1 D 4/ D 1 (8.212)
667
D 1 / C 1; 0.s
D 2 / C 3g (8.213)
D minf1; 1g D1
We observe that the result (8.213) is obtained for s 1 D 1 . Then the survivor sequence associated with s0 D 1 is L.s0 D 1 / D . 1/, expressed as a sequence of symbols rather than states. Considering now s0 D 2 , 3 , and 4 , in sequence, and applying (8.208), the rst iteration is completed. Obviously, there is no interest in determining the survivor sequence for states with metric equal to 1, because the corresponding path will not be extended. Next, for k D 1, the nal metrics and the survivor sequences are shown in the second diagram of Figure 8.24b, where we have 0.s1 D 1 / D minf1 C 1; 1g D 2 0.s1 D 2 / D minf1 C 2; 1g D 3 0.s1 D 3 / D minf1 C 1; 1g D 2 0.s1 D 4 / D minf1 C 0; 1g D 1 (8.214) (8.215) (8.216) (8.217)
The same procedure is repeated for k D 2; 3, and the trellis diagram is completed. The minimum among the values assumed by 0.s3 / is
i D1;:::;Ns
min
0.s3 D i / D 0.s3 D 2 / D 3
(8.218)
In practice, if the parameter K is large, the length of the survivor sequences is limited to a value K d , called trellis depth or path memory depth, that typically is between 3 and 10 times the length of the channel impulse response: this means that at every instant k we decide on ak K d CL 1 , a value that is then removed from the diagram. The decision is based on the survivor sequence associated with the minimum among the values of 0.sk /. In practice, k and 0.sk / may become very large; then the latter value is usually normalized by subtracting the same amount from all the metrics, for example the smallest of the metrics 0.sk D i /, i D 1; : : : ; Ns .
Computational complexity. The number of additions and comparisons is proportional to the number of transitions, M Ns D M .L 1 CL 2 C1/ . In any case, the complexity is linear in K .
668
We note that the values of f . j ; i /, that determine u k in (8.199) can be memorized in a table, limited to the possible transitions sk 1 D i ! sk D j , for i ; j 2 S. However, at every instant k and for every transition i ! j , the branch metric jk f . j ; i /j2 needs to be evaluated: this computation, however, can be done outside the recursive algorithm.
8.10.2
We now discuss an efcient method to compute the a posteriori probability (APP) in (8.185) for an isolated symbol. Given a vector z D [z 0 ; z 1 ; : : : ; z K 1 ]T , the notation zm indicates ` the vector formed only by the components [z ` ; z `C1 ; : : : ; z m ]T . We introduce the likelihood function, dened as
K Lk ./ D P[akCL 1 D j z0 1 K D 0 1
2A
(8.220)
k D 0; 1; : : : ; K
(8.221)
(8.222)
so that the desired signal at instant k, u k , given by (8.174), turns out to be only a function of sk and sk 1 (see (8.199)). The formulation that we give for the solution of the problem (8.221) is called a forwardbackward algorithm (FBA) and it follows the work by Rabiner [5]: it is seen that it coincides with the BCJR algorithm [6] for the decoding of convolutional codes. We observe that, unlike in (8.222), in the two formulations [5, 6] the denition of state also includes the symbol ak L 2 and consequently u k is only a function of the state at the instant k.
D i]
(8.223)
5. j j i/ D P[akCL 1 D ]
669
which represents the a priori probability of the generic symbol. There are algorithms to iteratively estimate this probability from the output of another decoder or equalizer (see Section 11.5). Here, for the time being, we assume there is no a priori knowledge on the symbols. Consequently, for i.i.d. symbols, we have that every state sk D j can be reached by M states, and P[akCL 1 D ] D If there is no transition from sk
1
1 M
2A
(8.225)
D i to sk D j , we set 5. j j i/ D 0 (8.226)
D i]
(8.227)
assuming that there is a transition from sk 1 D i to sk D j . For a channel with complex-valued additive Gaussian noise, (8.188) holds, and pzk .k j j; i/ D where u k D f . j ; i /. 4. We merge (8.223) and (8.227) by dening the variable C k . j j i/ D P[z k D k ; sk D j j sk D P[z k D k j sk D j ; sk D pzk .k j j; i/ 5. j j i/ 5. Initial and nal conditions are given by p j D P[s N
1 1 1
1 e 2 w
1 u j2 2 j w k k
(8.228)
D i] D i ] P[sk D j j sk
1
D i]
(8.229)
D j] D
1 Ns 1 Ns
j D 1; : : : ; Ns j D 1; : : : ; Ns i; j D 1; : : : ; Ns
q j D P[s K D j ] D N C K . j j i/ D 5. j j i/
If the initial and/or nal state are known, for example s we set pj D N
1
D i0 (
s K D f0
(8.233)
1 0
for j D i0 otherwise
(8.234)
670
(8.235)
(8.236)
Equation (8.236) gives the probability of observing the sequence 0 ; 1 ; : : : ; k up to instant k, and the state j at instant k. A recursive expression exists for Fk . j/. 1. Initialization F 1 . j/ D p j N 2. Updating for k D 0; 1; : : : ; K Fk . j/ D
Ns X `D1
j D 1; : : : ; Ns
(8.237)
1, C k . j j `/ Fk
1 .`/
j D 1; : : : ; Ns
(8.238)
Proof. Using the total probability theorem, and conditioning the event on the possible values of s k 1 , we express the probability in (8.236) as Fk . j/ D
Ns X `D1 Ns X `D1 k P[z0 1 k D 0 1
; z k D k ; s k D j ; s k
D `]
k P[z0
k D 0
; z k D k j s k D j ; s k
D ` ] P[s k D j ; s k
D `]
(8.239) Because the noise samples are i.i.d., once the values of s k and s k 1 are assigned, the event k k [z0 1 D 0 1 ] is independent of the event [z k D k ], and it results in Fk . j/ D
Ns X `D1 k P[z0 1 k D 0 1
j sk D j ; sk
1
D `]
1
(8.240) D `]
P[z k D k j s k D j ; s k
D ` ] P[s k D j ; s k
671
Moreover, given s k Fk . j/ D
Ns X `D1
1,
k D 0
k P[z0
j sk
D `]
1
(8.241)
1
D ` ] P[s k D j ; s k
D `]
; sk
D `]
1 P[s k
1
D `]
P[z k D k ; s k D j ; s k
D `]
k P[z0
k D 0
; sk
D ` ] P[z k D k ; s k D j j s k
D `] (8.242)
(8.243)
1,
Equation (8.243) is the probability of observing the sequence kC1 ; : : : ; K k C 1 onwards, given the state i at instant k. A recursive expression also exists for Bk .i/. 1. Initialization B K .i/ D qi N 2. Updating for k D K 1; K
Ns X mD1
from instant
i D 1; : : : ; Ns
(8.244)
Bk .i/ D
Proof. Using the total probability theorem, and conditioning the event on the possible values of s kC1 , we express the probability in (8.243) as Bk .i/ D
Ns X mD1 K K P[zkC11 D kC11 ; s kC1 D m j s k D i ]
(8.246) D
Ns X mD1 K K P[zkC11 D kC11 j s kC1 D m ; s k D i ] P[s kC1 D m j s k D i ]
672
K K Now, given the values of s kC1 and s k , the event [zkC21 D kC21 ] is independent of [z kC1 D K K kC1 ]. In turn, assigned the value of s kC1 , the event [zkC21 D kC21 ] is independent of s k . Then (8.246) becomes
Bk .i/ D
Ns X mD1
P[z kC1 D kC1 j s kC1 D m ; s k D i ] P[s kC1 D m j s k D i ] Observing (8.243) and (8.229), (8.245) follows. c) State metric
K Vk .i/ D P[sk D i j z0 1 K D 0 1
(8.248)
Equation (8.248) expresses the probability of being in the state i at instant k, given the K whole observation 0 1 . It can be expressed as a function of the forward and backward metrics, Vk .i/ D Fk .i/ Bk .i/ Ns X Fk .n/ Bk .n/
nD1
i D 1; : : : ; Ns
(8.249)
Proof. Using the fact that, given the value of s k , the r.v.s fz t g with t > k are statistically independent of fz t g with t k, from (8.248) it follows
K k k K Vk .i/ D P[z0 D 0 ; zkC11 D kC11 ; s k D i ]
1
K P[z0 1 K D 0 1
] 1 (8.250)
1 1 K D 0
k P[z0
k 0; sk
K i ] P[zkC11
K kC11
j sk D i ]
K P[z0
Observing the denitions of forward and backward metrics, (8.249) follows. We note that the normalization factor
K P[z0 1 K D 0 1 Ns X nD1
]D
Fk .n/ Bk .n/
(8.251)
Vk .i/ D 1
(8.252)
673
d) Likelihood function of the generic symbol. Applying the total probability theorem to (8.220) we obtain the relation Lk ./ D
Ns X i D1 K P[akCL 1 D ; sk D i j z0 1 K D 0 1
(8.253)
From the comparison of (8.253) with (8.248), indicating with [ i ]m , m D 1; : : : ; L 1 C L 2 , the m th component of the state i (see (8.222)), we have Lk ./ D
Ns X
i D1 condition [ i ]1 D
Vk .i/
2A
(8.254)
In other words, at instant k the likelihood function coincides with the sum of the metrics Vk .i/ associated with the states whose rst component is equal to the symbol of value . Note that Lk ./ can also be obtained using the state metrics evaluated at different instants, that is Lk ./ D
Ns X
i D1 [ i ]m D
VkC.m
1/ .i/
2A
(8.255)
for m 2 f1; : : : ; L 1 C L 2 g
Scaling
We see that, due to the exponential form of pzk .k j j; i/ in Ck . j j i/, in a few iterations the forward and backward metrics may assume very small values; this leads to numerical problems in the computation of the metrics: therefore we need to substitute equations (8.238) and (8.245) with analogous expressions that are scaled by a suitable coefcient. We note that the state metric (8.249) does not change if we multiply Fk .i/ and Bk .i/, i D 1; : : : ; Ns , by the same coefcient Kk . The idea [5] is to choose Kk D 1
Ns X nD1
(8.256)
Fk .n/
N C k . j j `/ Fk
1 .`/
(8.257)
674
N C k . j j `/ Fk
1 .`/
N Fk . j/ D
N C k .n j `/ Fk
1 .`/
j D 1; : : : ; Ns k D 0; 1; : : : ; K
(8.258)
j i/
1 .`/
N Bk .i/ D
i D 1; : : : ; Ns k D K 1; K
2; : : : ; 0
(8.259)
i D 1; : : : ; Ns k D 0; 1; : : : ; K
(8.260)
2A
(8.261)
e Vk ./ D X
2A
1 j2 2 j w k
2A
k D 0; 1; : : : ; K
(8.262)
675
(8.264) 1
B `k ./ D ln Lk ./ D ln B @
C evk .i / C A
2A
(8.265)
The function `k ./ is called log-likelihood. The exponential emphasizes the difference between the metrics vk .i/: typically a term dominates within each sum; this suggests the approximation ln
Ns X i D1
evk .i / '
i 2f1;:::;Ns g
max
vk .i/
(8.266)
max
vk .i/
(8.267)
and the Log-MAP criterion is replaced by the Max-Log-MAP criterion Q akCL 1 D arg max `k ./ O
2A
(8.268)
Observation 8.12 In the particular case of absence of ISI, substitution of (8.263) in (8.265) yields `k ./ D and (8.268) becomes ak D arg min jk O
2A
jk
2 w
j2
(8.269)
j2
(8.270)
which corresponds to the minimum distance decision criterion. Example 8.10.3 (Binary case) In the binary case, 2 f 1; 1g, it is convenient to introduce the likelihood ratio given by (see (8.220)) Lk D
K P[akCL 1 D 1 j z0 1 K D 0 1
P[akCL 1 D
K 1 j z0 1
K 0 1]
Lk .1/ Lk . 1/
(8.271)
676
1 1
if Lk 1 if Lk < 1
(8.272)
We recall that in the absence of ISI the expression of Lk ./ is given by (8.262). The analysis is simplied by the introduction of the log-likelihood ratio (LLR), `k D ln Lk D `k .1/ `k . 1/ (8.273)
where `k ./ is given by (8.265). Then (8.272) becomes akCL 1 D sgn.`k / O Observing (8.274), we can write `k D akCL 1 j`k j O (8.275) (8.274)
In other words, the sign of the log-likelihood ratio yields a hard-decision on the symbol being detected, while the magnitude indicates the reliability of the decision. The function `k is used as a soft-decision parameter in some detection algorithms (see page 923). In the Max-Log-MAP formulation, the LLR is given by Q Q `k D `k .1/ D max Q `k . 1/ vk .i/
i 2 f1; : : : ; Ns g [ i ]1 D 1
i 2 f1; : : : ; Ns g [ i ]1 D 1
max
vk .i/
(8.276)
Apart from non-essential constants, the Max-Log-MAP algorithm in the case of transmission of i.i.d. symbols over a channel with additive white Gaussian noise is formulated as follows. 1. Computation of channel transition metrics. For k D 0; 1; : : : ; K ck . j j i/ D jk u k j2 i; j D 1; : : : ; Ns
1
again, assuming there is a transition between i and j . 2. Backward procedure. For k D K Q bk .i/ D
m2f1;:::;N s g
1; K
2; : : : ; 0, i D 1; : : : ; Ns (8.280)
max
677
Q If the nal state is unknown, we set b K .i/ D 0, i D 1; : : : ; Ns . 3. Forward procedure. For k D 0; 1; : : : ; K fQk . j/ D
`2f1;:::;Ns g
1, j `/] j D 1; : : : ; Ns (8.282)
max
[ fQk
1 .`/ C ck . j
5. Log-likelihood function of an isolated symbol. For k D 0; 1; : : : ; K 1, the loglikelihood function is given by (8.267), with vk .i/ in substitution of vk .i/; the decision Q rule is given by (8.268). In practical implementations of the algorithm, steps 3, 4, and 5 can be carried out in sequence for each value of k: this saves memory locations. To avoid overow, for each Q value of k a common value can be added to all variables fQk .i/ and bk .i/, i D 1; : : : ; Ns . We observe that the two procedures, backward and forward, can be efciently implemented by the Viterbi algorithm, using a trellis diagram run both in backward and forward directions. The simplied MAP algorithm requires about twice the complexity of the VA implementing the MLSD criterion. Memory requirements are considerably increased with respect to the VA, because the backward metrics must be stored before evaluating the state metrics. However, methods for an efcient use of the memory are proposed in [7].
(8.285)
(8.286)
We now extend the above denition to the case of three variables, max .x; y; z/ D ln.e x C e y C e z / (8.287)
678
The extension to more variables is readily obtained by induction. So, if in the backward and forward procedures of page 676 we substitute the max function with the max function, we obtain the exact Log-MAP formulation that relates vk .i/ D ln Vk .i/ to bk .i/ D ln Bk .i/ and f k .i/ D ln Fk .i/, using the branch metric ck . j j i/.
8.11
It is possible to identify two different receiver structures that supply the signal z k given by (8.173). 1. The rst, illustrated in Figure 8.25a, is considered for the low implementation complexity; it refers to the receiver of Figure 7.12, where s f (8.289) ; GRc . f / D rcos 1=T and wC .t/ is white Gaussian noise with spectral density N0 . Recalling that r R .t/ D s R .t/ C w R .t/, with f (8.290) ; Pw R . f / D Pw . f / jGRc . f /j2 D N0 rcos 1=T
f (1/T ,) gRc
G R (f)= rcos
C
ak T
sC(t)
C
rC(t)
t 0+kT rR (t) zk T
wC (t)
(AWGN)
(a)
rC(t)
x(t)
M
t 0+kT WF xk zk w T T
wC (t) (AWGN)
(b)
Figure 8.25. Two receiver structures with i.i.d. noise samples at the decision point.
679
it is important to verify that the noise sequence fwk D w R .t0 C kT /g has a constant 2 spectral density equal to N0 , and the variance of the noise samples is w D N0 =T . Although the lter dened by (8.289) does not necessarily yield a sufcient statistic (see Observation 8.13 on page 681), it considerably reduces the noise and this may be useful in estimating the channel impulse response. Another problem concerns the optimum timing phase, which may be difcult to determine for non-minimum phase channels. 2. An alternative, also known as the Forney receiver, is represented in Figure 8.20 and repeated in Figure 8.25b. To construct the WF, however, it is necessary to determine poles and zeros of the function 8.z/; this can be rather complicated in real time applications. A practical method is based on Observation 8.9 on page 650. From the knowledge of the autocorrelation sequence of the channel impulse response, the prediction error lter A.z/ is determined. The WF of Figure 8.25 is given by W .z/ D 12 A .1=z /; therefore it is an FIR lter. The impulse response fn g is given by the inverse z-transform of 1 f0 F.z/ D 1=A.z/. For the realization of symbol detection algorithms, a windowed version of the impulse response is considered. A further method, which usually requires a lter w with a smaller number of coefcients than the previous method, is based on the observation that, for channels with low noise level, the DFE solution determined by the MSE method coincides with the DFE-ZF solution; in this case the FF lter plays the role of the lter w in Figure 8.25b. We consider two cases. a. At the output of the MF g M , let fh n g be the system impulse response with sampling period T , determined for example through the method described in the Observation 8.5 on page 641. Using (8.96), a DFE is designed with lter parameters .M1 ; M2 /, and consequently w D copt in Figure 8.25b. At the output of the lter w, the ideally minimum phase impulse response fn g corresponds to the translated, by D sampling intervals, and windowed, with L 1 D 0 and L 2 D M2 , version of D h copt . b. If the impulse response of the system is unknown, we can resort to the FSDFE structure of Figure 8.14. Using now an adaptive method to determine the coefcients of the DFE, at convergence we get ( 1 nD0 n D nCD ' (8.291) bn n D 1; : : : ; M2 Actually, unless the length of fn g is shorter than that of the impulse response qC , it is convenient to use Ungerboecks formulation of the MLSD [8] that utilizes only samples fx k g at the MF output; now, however, the metric is no longer Euclidean. The derivation of the non-Euclidean metric is the subject of the next section. We note that, as it is not important to obtain the likelihood of an isolated symbol from the non-Euclidean metric, there are cases in which this method is not adequate. We refer in particular to the case in which decoding with soft input is performed separately from
f0
680
symbol detection in the presence of ISI (see Section 11.3.2). However, joint decoding and detection are always possible using a suitable trellis (see Section 11.3.2).
(8.292)
ak qC .t
kT /
(8.293)
For a suitable basis, we consider for (8.292) the following vector representation: rDsCw Assuming that the transmitted symbol sequence a D [a0 ; a1 ; : : : ; a K 1 ] is equal to D [0 ; 1 ; : : : ; K 1 ], the probability density function of r is given by (6.16), 1 2 (8.294) prja . j / D K exp jj sjj N0 Using (1.36), and observing rC .t/ D .t/, we get Z 1 prja . j / D K exp j.t/ N 0 TK (8.295)
sC .t/j2 dt
Taking the logarithm in (8.295), the log-likelihood (to be maximized) is 2 Z K 1 X `./ D k qC .t kT / dt .t/ TK kD0 Correspondingly the detected sequence is given by O ak D arg max `./
(8.296)
(8.297)
Expanding the squared term in (8.296), we obtain (Z "Z K 1 X 2 `./ D j.t/j dt 2Re .t/ k qC .t
TK TK kD0
# kT / dt ) (8.298)
Z C
K 1 K 1 X X TK k1 D0 k2 D0
k1 k2 qC .t
k1 T / qC .t
k2 T / dt
681
t/
(8.299)
and the overall impulse response at the MF output q.t/ D .qC g M /.t/ D rqC .t t0 / (8.300)
where rqC is the autocorrelation of qC , whose samples are given by (see (8.23)) h n D q.t0 C nT / D rqC .nT / Let x.t/ be the MF output signal expressed as x.t/ D .rC g M /.t/ with samples given by x k D x.t0 C kT / (8.303) (8.302) (8.301)
In (8.298) the rst term can be ignored since it does not depend on , while the other two terms are rewritten in the following form: ( " # ) K 1 K 1K 1 X X X `./ D 2Re k x k C k 1 k 2 h k 2 k 1 (8.304)
kD0 k1 D0 k2 D0
Observing (8.304), we obtain the following important result. Observation 8.13 The sequence of samples fx k g, taken by sampling the MF output signal with sampling period equal to the symbol period T , forms a sufcient statistic to detect the message fak g associated with the signal rC dened in (8.292). We express the double summation in (8.304) as the sum of three terms, the rst for k1 D k2 , the second for k1 < k2 , and the third for k1 > k2 : AD
K 1K 1 X X k1 D0 k2 D0 k 1 k 2 h k 2 k1
(8.305) D
K 1 X k1 D0 k 1 k 1 h 0 C K 1 k1 1 X X k1 D1 k2 D0 k 1 k 2 h k 2 k1
K 1 k2 1 X X k2 D1 k1 D0
k 1 k 2 h k 2
k1
Because the sequence fh n g is an autocorrelation, it enjoys the Hermitian property, i.e. h n D h n ; consequently, the third term in (8.305) is the complex conjugate of the second, and " # K 1 K 1X X X k 1 AD k k h 0 C 2Re k k 2 h k k 2 (8.306)
kD0 kD1 k2 D0
682
( D Re In particular, if
K 1 X kD0
"
k h 0 k C 2
k X nD1
#) h n k
n
(8.307)
jh n j ' 0
(8.308)
(8.309)
#)
n
(8.310)
To maximize `./ or, equivalently, to minimize `./ with respect to we apply the Viterbi algorithm (see page 663) with the state vector dened as sk D .ak ; ak and branch metric given by ( Re
ak 1 ; : : : ; ak Nh C1 /
(8.311) #)
" 2x k C h 0 ak C 2
Nh X nD1
h n ak
(8.312)
Extensions of Ungerboecks approach to time variant radio channels are proposed in [9].
8.12
In the Viterbi algorithm, we have an error if a sequence of states that is different from the correct sequence is chosen as maximum likelihood sequence in the trellis diagram; the probability that one or more states of the detected ML sequence are in error is interesting. The error probability is dominated by the probability that a sequence at the minimum Euclidean distance from the correct sequence is chosen as ML sequence. We note, however, that increasing the sequence length K also increases the number of different paths in the trellis diagram associated with sequences that are at the minimum distance. Therefore, by increasing K , the probability that the chosen sequence is in error usually tends to 1. The probability that the whole sequence of states is not received correctly is rarely of interest; instead, we consider the probability that the detection of a generic symbol is in
683
error. For the purpose of determining the symbol error probability, the concept of error event is introduced. Let f g D . i0 ; : : : ; i K 1 / be the realization of the state sequence associated with the information sequence, and let f g be the sequence chosen by the Viterbi algorithm. O In a sufciently long time interval, the paths in the trellis diagram associated with f g and f g diverge and converge several times: every distinct separation from the correct path is O called an error event. Denition 8.1 An error event e is dened as a path in the trellis diagram that has only the initial and nal states in common with the correct path; the length of an error event is equal to the number of nodes visited in the trellis before rejoining with the correct path. Error events of length one and two are illustrated in a trellis diagram with two states, where the correct path is represented by a continuous line, in Figure 8.26a and Figure 8.26b, respectively. Let E be the set of all error events beginning at instant i. Each element e of E is characterized by a correct path f g and a wrong path f g, which diverges from f g at O instant i and converges at f g after a certain number of steps in the trellis diagram. We assume that the probability P[e] is independent of instant i: this hypothesis is veried with good approximation if the length of the trellis diagram is much greater than the length of the signicant error events. An error event produces one or more errors in the detection of symbols of the input sequence. We have a detection error at instant k if the detection of the input at the k-th stage of the trellis diagram is not correct. We dene the function [10] ( cm .e/ D 1 0 if e causes a detection error at the instant i C m ; with m otherwise 0
(8.313) The probability of a particular error event that starts at instant i and causes a detection error at instant k is given by ck i .e/P[e]. Because the error events in E are disjointed, we have Pe D P[ak 6D ak ] D O
k X X i D 1 e2E
ck i .e/ P[e]
(8.314)
k+1
k+2
k+1
k+2
k+3
(a)
(b)
Figure 8.26. Error events of length (a) one and (b) two in a trellis diagram with two states.
684
P[e]
k X iD 1
ck i .e/
(8.315)
ck i .e/ D
1 X mD0
cm .e/ D N .e/
(8.316)
which indicates the total number of detection errors caused by the error event e. Therefore, X Pe D N .e/P[e] (8.317)
e2E
where the dependence on the time index k vanishes. We therefore nd that the detection error probability is equal to the average number of errors caused by all the possible error events initiating at a given instant i; this result is expected, because the detection error probability at a particular instant k must take into consideration all error events that initiate at previous instants and are not yet terminated. If fsg D .s0 ; : : : ; s K 1 / denotes the random variable sequence of states at the transmitter and fO g D .O 0 ; : : : ; s K 1 / denotes the random variable sequence of states selected s s O by the ML receiver, the probability of an error event e beginning at a given instant i depends on the joint probability of the correct and incorrect path, and it can be written as P[e] D P[fO g D f g j fsg D f g]P[fsg D f g] s O (8.318)
Because it is usually difcult to nd the exact expression for P[fO g D f g j fsg D f g], we s O resort to upper and lower limits. Upper limit. Because detection of the sequence of states fsg is obtained by observing the sequence fug; for the signal in (8.173) with zero mean additive white Gaussian noise having variance I2 per dimension, we have the upper limit d[u.f g/; u.f g/] O P[fO g D f g j fsg D f g] Q s O (8.319) 2 I where d[u.f g/; u.f g/] is the Euclidean distance between signals u.f g/ and u.f g/, given O O by (8.193). Substitution of the upper limit in (8.317) yields X d[u.f g/; u.f g/] O Pe (8.320) N .e/ P[fsg D f g]Q 2 I e2E which can be rewritten as follows, by giving prominence to the more signicant terms, X dmin C other terms (8.321) Pe N .e/ P[fsg D f g]Q 2 I e2Emin
685
where Emin is the set of error events at minimum distance dmin dened in (8.194), and the remaining terms are characterized by arguments of the Q function larger than dmin =.2 I /. For higher values of the signal-to-noise ratio these terms are negligible and the following approximation holds dmin (8.322) Pe K1 Q 2 I where K1 D X
e2Emin
N .e/ P[fsg D f g]
(8.323)
Lower limit. A lower limit to the error probability is obtained by considering the probability that any error event may occur rather than the probability of a particular error event. Since N .e/ 1 for all the error events e, from (8.317) we have X Pe P[e] (8.324)
e2E
Let us consider a particular path in the trellis diagram determined by the sequence of states f g. We set Q dmin .f g/ D min d[u.f g/; u.f g/]
f g Q
(8.325)
i.e., for this path, dmin .f g/ is the Euclidean distance of the minimum distance error event. We have dmin .f g/ dmin , where dmin is the minimum distance obtained considering all the possible state sequences. If f g is the correct state sequence, the probability of an error event is lower limited by dmin .f g/ (8.326) P[e j fsg D f g] Q 2 I Consequently, Pe X
f g
P[fsg D f g]Q
dmin .f g/ 2 I
(8.327)
If some terms are omitted in the equation, the lower limit is still valid, because the terms are non-negative. Therefore, taking into consideration only those state sequences f g for which dmin .f g/ D dmin , we obtain X dmin (8.328) Pe P[fsg D f g]Q 2 I f g2A where A is the set of state sequences that admit an error event with minimum distance dmin , for an arbitrarily chosen initial instant of the given error event. Dening X K2 D P[fsg D f g] (8.329)
f g2A
686
as the probability that a path f g admits an error event with minimum distance, it is dmin (8.330) Pe K2 Q 2 I Combining upper and lower limits we obtain dmin dmin K2 Q Pe K1 Q 2 I 2 I For large values of the signal-to-noise ratio, therefore we have dmin Pe ' K Q 2 I
(8.331)
(8.332)
for some value of the constant K between K1 and K2 . We stress that the error probability, expressed by (8.332) and (8.195), is determined by the ratio between the minimum distance dmin and the standard deviation of the noise I . Here the expressions of the constants K1 and K2 are obtained by resorting to various approximations. An accurate method to calculate upper and lower limits of the error probability is proposed in [11].
n ak
(8.333)
and variance I2 per dimension. The metric that the Viterbi algorithm attributes to the sequence of states corresponding to the sequence of input symbols fak g is given by the squared Euclidean distance between the sequence of samples fz k g at the detector input and its mean value, which is known, given the sequence of symbols (see (8.189)), 1 X z k kD0
L2 X nD L 1
2 n ak n
(8.334)
In the previous section it was demonstrated that the symbol error probability is given by (8.332). In particularly simple cases, the minimum distance can be determined by direct inspection of the trellis diagram; in practice, however, this situation is rarely veried in channels with ISI. To evaluate the minimum distance it is necessary to resort to simulations. To nd the minimum distance error event with initial instant k D 0, we consider the
687
desired signal u k under the condition that the sequence fak g is transmitted, and we compute the squared Euclidean distance between this signal and the signal obtained for another sequence fak g, Q L2 1 X X d [u.fak g/; u.fak g/] D Q n ak kD0 nD L
2
1
L2 X nD L 1
2 n ak n Q
(8.335)
where it is assumed that the two paths identifying the state sequences are identical for k < 0. It is possible to avoid computing the minimum distance for each sequence fak g if we exploit the linearity of the ISI. Dening k D ak we have L2 1 X X d .fk g/ D d [u.fak g/; u.fak g/] D Q n k kD0 nD L
2 2
1
ak Q 2 n
(8.336)
(8.337)
The minimum among the squared Euclidean distances relative to all error events that initiate at k D 0 is
2 dmin D fk g: k D0; k<L 1 ; L 1 6D0
min
d 2 .fk g/
(8.338)
It is convenient to solve this minimization without referring to the symbol sequences. In particular, we dene the state sk D .kCL 1 ; kCL 1 1 ; : : : ; k L 2 C1 /, and a trellis diagram that describes the development of this state. Adopting the branch metric 2 L2 X n k n (8.339) nD L
1
the minimization problem is equivalent to determining the path in a trellis diagram that has minimum metric (8.337) and differs from the path that joins states corresponding to correct 2 decisions: the resulting metric is dmin . We note, however, that the cardinality of k is larger than M, and this implies that the complexity of this trellis diagram can be much larger than that of the original trellis diagram. In the PAM case, the cardinality of k is equal to 2M 1. In practice, as the terms of the series in (8.337) are non-negative, if we truncate the series after a nite number of terms we obtain a result that is smaller than or equal to the effective value. Therefore a lower limit to the minimum distance is given by 2 K 1 X X L2 2 min n k n (8.340) dmin nD L fk g: k D0; k<L 1 ; L 1 6D0 kD0
1
Example 8.12.1 We consider the partial response system class IV (PR-IV), also known as modied duobinary (see Appendix 7.A), illustrated in Figure 8.27. The transfer function of the discrete-time
688
bk ak u k = a k a k2
Figure 8.28. Input and output sequences for an ideal PR-IV system.
overall system is given by .D/ D 1 D 2 . For an ideal noiseless system, the input sequence fu k g to the detector is formed by random variables taking values in the set f 2; 0; C2g, as shown in Figure 8.28. Assuming that the sequence of noise samples fwk g is composed of real-valued, statistically independent, Gaussian random variables with mean zero and variance I2 , and observing that u k for k even (odd) depends only on symbols with even (odd) indices, the MLSD receiver for a PR-IV system is usually implemented by considering two interlaced dicode independent channels, each having a transfer function given by 1 D. As seen in Figure 8.29, for detection of the two interlaced input symbol sequences, two trellis diagrams are used. The state at instant k is given by the symbol sk D ak , where k is
a k*2 +1 bk 1 0 1 1 0 uk 0 +2 2 0 b k)2 u k)2 1 0 1 0 0 +2 2 0 b k)4 u k)4 1 0 1 0 0 +2 2 0 b k)6 u k)6 1 0 1 0 0 +2 2 0
a k*1 +1
bk
BMAP 1 a k 0 +1 T 1
wk (D) uk zk MLSD ^ ak
... 0
... 1 +1 +1 1 +1 1 ... +2 2 0 0
b k)1 u k)1 1 0 1 0 0 +2 2 0
b k)3 u k)3 1 0 1 0 0 +2 2 0
b k)5 u k)5 1 0 1 0 0 +2 2 0
b k)7 u k)7 1 0 1 0 0 +2 2 0
689
k= 0 1 =1 2 =1 b0 =0
b0 =0
b1 =0
b0 =0
b1 =0 b 2 =1 b3 =1
b0 =0
b1 =0 b 2 =1 b3 =1
b0 =0
b1 =0
Figure 8.30. Survivor sequences at successive iterations of the Viterbi algorithm for a dicode channel.
even-valued in one of the two diagrams and odd-valued in the other. Each branch of the diagram is marked with a label that represents the binary input symbol bk or, equivalently, the value of the dicode signal u k . For a particular realization of the output signal of a dicode channel, the two survivor sequences at successive iterations of the Viterbi algorithm are represented in Figure 8.30. It is seen that the minimum squared Euclidean distance between two separate paths in 2 the trellis diagram is given by dmin D 22 C 22 D 8. However, we note that for the same initial instant, there are an innite number of error events with minimum distance from the effective path: this fact is evident in the trellis diagram of Figure 8.31a, where the state sk D .k / 2 f 2; 0; C2g characterizes the development of the error event, and the labels indicate the branch metrics associated with an error event. It is seen that at every instant an error event may be extended along a path, for which the metric is equal to zero, parallel
690
sk 2 4 16 0 16 4 +2 0 (a) 4 0 (b) 16 0 4 16 4 4 4 0
Figure 8.31. Examples of (a) trellis diagram to compute the minimum distance for a dicode channel and (b) four error events with minimum distance.
to the path corresponding to the zero sequence. Paths of this type correspond to a sequence of errors having the same polarity. Four error events with minimum distance are shown in Figure 8.31b. p The error probability is given by KQ 28 , where K2 K K1 . The constant K2 I can be immediately determined by noting that every effective path admits at every instant at least one error event with minimum distance: consequently K2 D 1. To nd K1 , we consider the contribution of an error event with m consecutive errors. For this event to occur, it is required that m consecutive input symbols have the same polarity, which happens with probability 2 m . Since such an error event determines m symbol errors and two error events with identical characteristics can be identied in the trellis diagram, we have K1 D 2
1 X mD1
m2
D4
(8.341)
p Besides the error events associated with the minimum distance dmin D 8, the occurrence of long sequences of identical output symbols from a dicode channel raises two problems. The system becomes catastrophic in the sense that in the trellis diagram used by the detector, valid sequences of arbitrary length are found having squared Euclidean distance from the effective path equal to 4; therefore, a MLSD receiver with p nite memory will make additional errors with probability proportional to Q 24 if I the channel produces sequences fu k D 0g of length larger than the memory of the detector. The occurrence of these sequences is detrimental for the control of receive lter gain and sampling instants: the problem is solved by suitable coding of the input binary sequence, that sets a limit on the number of consecutive identical symbols that are allowed at the channel input.
691
8.13
For transmission over channels characterized by strong ISI, maximum likelihood detection implemented by the Viterbi algorithm typically gives better performance than a decisionfeedback equalizer. On the other hand, if we indicate with N D L 1 C L 2 the length of the channel memory and with M the cardinality of the alphabet of transmitted symbols, the implementation complexity of the VA is of the order of M N , which can be too large for practical applications if N and/or M assume large values. To nd receiver structures that are characterized by performance similar to that of a MLSD receiver, but lower complexity, two directions may be taken: decrease N ; decrease the number of paths considered in the trellis; The rst direction leads to the application of pre-processing techniques, e.g., LE or DFE [12], to reduce the length of the channel impulse response. However, they only partially solve the problem: for bandwidth efcient modulation systems that utilize a large set of signals, the minimization of N is often insufcient to signicantly reduce the complexity. The second direction leads to further study of the MLSD method. An interesting algorithm is the M-algorithm [13, 14]: it operates in the same way as the MLSD, i.e. using the full trellis diagram, but at every step it takes into account only M M N states, that is, those associated with paths with smaller metrics. The performance of the M-algorithm is close to that of the MLSD even for small values of M, however, it works only if the channel impulse response is minimum phase: otherwise, it is likely that among the paths that are eliminated in the trellis diagram the optimum path is also included. The reduced state sequence estimator (RSSE)4 [15, 16] is capable of yielding performance very close to that of MLSD, with signicantly reduced complexity even though it retains the fundamental structure of MLSD. It can be used for modulation with a large symbol alphabet A and/or for channels with a long impulse response. The basic idea consists in using a trellis diagram with a reduced number of states, obtained by combining the states of the ML trellis diagram in a manner suggested by the principle of partitioning the set A (see Chapter 12), and possibly including the decision-feedback method in the computation of the branch metrics. In this way the RSSE guarantees a performance/complexity trade-off that can vary from that characterizing a DFE zero-forcing to that characterizing MLSD.
We will maintain the name RSSE, although the algorithm is applied to perform a detection rather than an estimation.
692
response that determine the ISI and assume, without loss of generality, the desired sample is f 0 D 1. Hence the observed signal is given by zk D where the state at instant k
N X nD0
f n ak
C wk D ak C hsk
1 ; fi C wk
(8.342)
1 is given by
T sk 1
D [ak
1 ; ak 2 ; : : : ; a k N ]
(8.343)
and fT D [ f 1 ; f 2 ; : : : ; f N ] (8.344)
Observation 8.14 In the transmission of sequences of blocks of data, the RSSE yields best performance by imposing the nal state, for example, using the knowledge of a training sequence. Therefore the formulation of this section is suited for the case of a training sequence placed at the end of the data block. In the case the training sequence is placed at the beginning of a data block, it is better to process the signals in backward mode as described in Observation 8.6 on page 642. The RSSE maintains the fundamental structure of MLSD unaltered, corresponding to the search in the trellis diagram of the path with minimum cost. To reduce the number of states to be considered, we introduce for every component ak n , n D 1; : : : ; N , of the vector sk 1 dened in (8.343) a suitable partition .n/ of the two-dimensional set A of possible values of ak n : a partition is composed of Jn subsets, with Jn an integer between 1 and M. The index of the subset of the partition .n/ to which the symbol ak n belongs is indicated by cn , an integer value between 0 and Jn 1. Ungerboecks partitioning of the symbol set associated with a 16-QAM system is illustrated in Figure 8.32. The partitions must satisfy the following two conditions (see also Chapter 12, or, for a more general partitioning method, [17]): 1. the numbers Jn are non-increasing, that is J1 J2 J N ; 2. the partition .n/ is obtained by subdividing the subsets that make up the partition .n C 1/, for every n between 1 and N 1. Therefore we dene as reduced state at instant k 1 the vector tk 1 that has as n-th element the index cn of the subset of the partition .n/ to which the n-th element of sk 1 belongs, for n D 1; 2; : : : ; N , that is
T tk 1
D [c1 ; c2 ; : : : ; c N ]
cn 2 f0; 1; : : : ; Jn
1g
(8.345)
and we write sk
1
D s.tk
1/
(8.346)
693
}J
0 1
}J
0 2 1 3
}J
0 4 2 6 1 5 3 7
}J
0 8 4 12 2 10 6 14 1 9 5 13 3 11 7 15
}J
16
Figure 8.32. Ungerboecks partitioning of the symbol set associated with a 16-QAM system. The various subsets are identied by the value of cn 2 f0; 1; : : : ; Jn 1g.
It is useful to stress that the reduced state tk 1 does not uniquely identify a state sk 1 , but all the states sk 1 that include as n-th element one of the symbols belonging to the subset cn of partition .n/. The conditions imposed on the partitions guarantee that, given a reduced state at instant k 1, tk 1 , and the subset j of partition .1/ to which the symbol ak belongs, the reduced state at instant k, tk , can be uniquely determined. In fact, observing (8.345), we have
T 0 0 tk D [c1 ; c2 ; : : : ; c0N ]
(8.347)
0 0 where c1 D j, c2 is the index of the subset of the partition .2/ to which belongs the 0 subset with index c1 of the partition .1/, c3 is the index of the subset of the partition .3/ to which belongs the subset with index c2 of the partition .2/, and so forth. In this way the reduced states tk 1 dene a proper reduced state trellis diagram, that represents all the possible sequences fak g. As the symbol cn can only assume one of the integer values between 0 and Jn 1, the total number of possible reduced states of the trellis diagram of the RSSE is given by the product Ns D J1 J2 : : : J N , with Jn M, for n D 1; 2; : : : ; N . We know that in the VA, for uncoded transmission of i.i.d. symbols, there are M possible transitions from a state, one for each of the values that ak can assume. In the reduced state trellis diagram M transitions are still possible from a state, however, to only J1 distinct
694
states, thus giving origin to parallel transitions.5 In fact, if J1 < M, J1 sets of branches depart from every state tk 1 , each set consisting of as many parallel transitions as there are symbols belonging to the subset of .1/ associated with the reduced state. Therefore partitions must be obtained such that two effects are guaranteed: 1) minimum performance degradation with respect to MLSD, and 2) easy search of the optimum path among the various parallel transitions. The method that is usually adopted is Ungerboecks set partitioning method, which, for every partition , maximizes the minimum distance 1 among the symbols belonging to the same subset. For QAM systems and Jn a power of 2, the maximum distance 1n relative to partition .n/ is obtained through a tree diagram with binary partitioning (see Chapter 12). An example of partitioning of the symbol set associated with a 16-QAM system is illustrated in Figure 8.32. In Figure 8.33 two examples of reduced state trellis diagram are shown, both referring to the partition of Figure 8.32.
RSSE algorithm
As in MLSD, for each transition that originates from a state tk branch metric according to the expression
t k-1
[0, 0]
tk
[0, 0]
t k-1
[0]
tk
[0 ]
[0, 1]
[0, 1]
[1, 0]
[1, 0]
[1]
[1]
[1, 1]
[1, 1]
[2]
[2]
[2, 0]
[2, 0]
[2, 1]
[3, 0]
[3, 0]
[3, 1]
[3, 1]
(a) N D 1, J1 D 4.
(b) N D 2, J1 D 4, J2 D 2.
Parallel transitions are present when two or more branches connect the same pair of states in the trellis diagram.
695
jz k
ak
hsk
2 1 ; fij
D jz k
ak
hs.tk
2 1 /; fij
(8.348)
However, whereas in MLSD a survivor sequence may be described in terms of the sequence of states that led to a certain state, in the RSSE, if J1 < M, it is convenient to memorize a survivor sequence as a sequence of symbols. As a matter of fact there is no one-to-one correspondence between state sequences and symbol sequences. Therefore backward tracing the optimum path in the trellis diagram in terms of the sequence of states does not univocally establish the optimum sequence of symbols. Moreover, if J1 < M, and therefore there are parallel transitions in the trellis diagram, for every branch the RSSE selects the symbol ak in the subset of the partition .1/ that yields the minimum metric.6 Thus the RSSE already makes a decision in selecting one of the parallel transitions, using also the past decisions memorized in the survivor sequence associated with the considered state. At every iteration these decisions reduce the number of possible extensions of the Ns states from Ns M to Ns J1 . This number is further reduced to Ns by selecting, for each state tk , the path with the minimum metric among the J1 entering paths, operation which requires Ns .J1 1/ comparisons. As in the case of the VA, nal decisions are taken with a certain delay by tracing the history of the path with lowest metric. We note that if Jn D 1, n D 1; : : : ; N , then the RSSE becomes a DFE-ZF, and if Jn D M, n D 1; : : : ; N , the RSSE performs full MLSD. Therefore the choice of fJn g determines a trade-off between performance and computational complexity. The error probability for an RSSE is rather difcult to evaluate because of the presence of the decision-feedback mechanism. For the analysis of RSSE performance we refer the reader to [18].
We note that if the points of the subsets of the partition .1/ are on a rectangular grid, as in the example of Figure 8.32, the value of the symbol ak that minimizes the metric is determined through simple quantization rules, without explicitly evaluating the branch metric for every symbol of the subset. Hence, for every state, only J1 explicit computations of the branch metric are needed.
696
a. The estimator associated with the survivor sequence uses symbols that can be considered decisions with no delay and high reliability, making the PSP a suitable approach for channels that are fast time-varying. b. Blind techniques, i.e. without knowledge of the training sequence, may be adopted as part of the PSP to estimate the various parameters. A variant of the RSSE, belonging to the class of PSP algorithms, is the decision feedback sequence estimator (DFSE) [15] which considers as reduced state the vector s0 1 formed k simply by truncating the ML state vector at a length N1 N sk
0T
D [ak
1 ; ak 2 ; : : : ; a k N 1 ]
(8.349)
This is the same as considering Jn D M, for n D 1; : : : ; N1 , and Jn D 1, for n D N1 C 1; : : : ; N . We discuss the main points of this algorithm. We express the received sequence fz k g, always under the hypothesis of a minimum-phase overall impulse response with f 0 D 1, as z k D ak C hs 0 k where f0 and f00 , s0 k
1, 1; f 0
i C hs 00 1 ; f00 i C wk k
(8.350)
and s00 k
fT D [f0 T j f00 T ] D [ f 1 ; : : : ; f N1 j f N1 C1 ; : : : ; f N ]
T sk 1
D [sk
0T
j sk
00 T
1]
D [ak
1 ; : : : ; ak N 1
j ak
.N1 C1/ ; : : : ; ak N ]
The trellis diagram is built by assuming the reduced state s0 1 . The term hs00 1 ; f00 i reprek k sents the residual ISI that is estimated by considering as s00 1 the symbols that are memok rized in the survivor sequence associated with each state. We write s 00 Ok
1
D s00 .s0 k
1/
(8.353)
In fact, with respect to z k , it is as if we have cancelled the term hs00 1 ; f00 i by a FB lter k associated with each state s0 1 . With respect to the optimum path, the feedback sequence k s00 1 is expected to be very reliable. k The branch metric of the DFSE is computed as follows: jz k ak hs0 k
1; f 0
hs00 .s0 k
1 /; f
00 2
ij
(8.354)
We note that the reduced state s0 1 may be further reduced by adopting the RSSE techk nique (8.346). The primary difference between an MLSD receiver and the DFSE is that in the trellis diagram used by the DFSE two paths may merge earlier, as it is sufcient that they share the
697
more recent N1 symbols, rather than N as in MLSD. This increases the error probability; however, the performance of the DFSE is better than that achieved by a classical DFE.
8.14
Passband equalizers
For QAM signals, we analyze alternatives to the baseband equalizers considered at the beginning of this chapter. We refer to the QAM transmitter scheme of Figure 7.5 and we consider the model of Figure 8.34, where the transmission channel has impulse response gCh and introduces additive white noise w.t/. The received passband signal is given by " r.t/ D Re
C1 X kD 1
ak qCh .t
kT /e
j .2 f 0 tC'/
# C w.t/ (8.355)
where, from (7.42) or equivalently from Figure 7.11, qCh is a baseband equivalent pulse with frequency response
.bb/ QCh . f / D HTx . f / 1 GCh . f / D HTx . f / GCh . f C f 0 / 1. f C f 0 / 2
(8.356)
In this model the phase ' in (8.355) implies that arg QCh .0/ D 0, hence it includes also the phase offset introduced by the channel frequency response at f D f 0 , in addition to the carrier phase offset between the transmit and receive carriers. Because of this impairment, which as a rst approximation is equivalent to a rotation of the symbol constellation, a suitable receiver structure needs to be developed [10]. As ' may be time-varying, it can be decomposed as the sum of three terms '.t/ D 1' C 2 1 f t C .t/ (8.357)
where 1' is a xed phase offset, 1 f is a xed frequency offset, and .t/ is a random or quasi-periodic term (see the denition of phase noise in (4.271)). For example, over telephone channels typically j .t/j =20. Moreover, the highest frequency of the spectral components of .t/ is usually lower than 0:1=T : in other words, if .t/ were a sinusoidal signal, it would have a period larger than 10T . Therefore '.t/ may be regarded as a constant, or at least as slowly time varying, at least for a time interval equal to the duration of the overall system impulse response.
Figure 8.34. Passband modulation scheme with phase offset introduced by the channel.
698
8.14.1
Filtering and equalization of the signal r are performed in the passband. First, a lter extracts the positive frequency components of r, acting also as a matched lter. It is a complex-valued passband lter, as illustrated in Figure 8.35. In particular, from the theory of the optimum receiver we have g M .t/ D g M .t/e j2 f 0 t where
g M .t/ D qCh .t0 . pb/
(8.358)
t/
(8.359)
with qCh dened in (8.356). Then, with reference to Figure 8.36, we have g M;I .t/ D Refg M .t/g and
. pb/ . pb/ . pb/
. pb/
. pb/
(8.360)
If g M in (8.359) is real-valued, then g M;I and g M;Q are related by the Hilbert transform (1.163), that is g M;Q .t/ D H.h/ [g M;I .t/]
. pb/ . pb/
(8.361)
699
After the passband matched lter, the signal is oversampled with sampling period T =F0 ; oversampling is suggested by the following two reasons: 1. If qCh is unknown and g M is a simple passband lter, matched ltering is carried out by the lter c. pb/ , hence the need for oversampling. 2. If the timing phase t0 is not accurate, it is convenient to use an FSE. Let q.t/ be the overall baseband equivalent impulse response of the system at the sampler input, q.t/ D F where Q. f / D QCh . f / G M . f / D HTx . f / GCh . f C f 0 / G M . f / 1. f C f 0 / The sampled passband signal is given by T T . pb/ . pb/ . pb/ t0 C n C j xQ t0 C n xn D x I F0 F0 D where7
C1 X kD 1 1 . pb/
[Q. f /]
(8.362)
(8.363)
ak h n
k F0 e
T j 2 f 0 n F C'
0
(8.364) C wn Q
T h n D q t0 C n F0
(8.365)
(8.366)
where w R .t/ D w g M .t/. For a passband equalizer with N coefcients, the output signal with sampling period equal to T is given by yk
. pb/
N 1 X i D0
ci
D xk F0 c. pb/
. pb/T
(8.367)
with the usual meaning of the two vectors xn Ideally it should result yk
. pb/
D ak
De
j .2 f 0 kT C'/
700
In Figure 8.36 the signal yk is shifted to baseband by multiplication with the function j .2 f 0 kT C'/ , where ' is an estimate of '. Then the data detector follows. O e O At this point some observations can be made: by demodulating the received signal, that is by multiplying it with the function e j2 f 0 t , before the equalizer or the receive lter we obtain a scheme equivalent to that of Figure 8.3, with a baseband equalizer. As we will see at the end of this section, the only advantage of a passband equalizer is that the computational complexity of the receiver is reduced; in any case, it is desirable to compensate for the presence of the phase offset, that is to multiply the received signal by O e j ' , as near as possible to the decision point, so that the delay in the loop for the update of the phase offset estimate is small.
. pb/
yk
(8.369)
where yk is given by (8.367), we desire to minimize the following cost function: J D E[jek j2 ] D E[jak D E[jak
D De
xk
. pb/T . pb/
j .2 f 0 kT C'/ 2 O
j ]
(8.370)
j .2 f 0 kT C'/ O
xk
. pb/T . pb/ 2
j ]
Equation (8.370) expresses the classical Wiener problem for a desired signal expressed as ak and input xk
. pb/ De j .2 f 0 kT C'/ O
(8.371)
(8.372)
(8.373)
m/T /
m/T
.`
C rw R ..`
m/T /
(8.374)
and p D E[ak
De j .2 f 0 kT C'/ . pb/ O xk ]
(8.375)
701
(8.376)
where
2 [p0 ]` D a h D `e j2 f 0 `T
(8.377)
pe
j .' '/ O
(8.378)
where R 1 and p0 do not depend on the phase offset '. From (8.378) it can be veried that if ' is a constant the equalizer automatically com. pb/ remains pensates for the phase offset introduced by the channel, and the output signal yk unchanged. A difculty appears if ' varies (slowly) in time and the equalizer attempts to track it. In fact, to avoid the output signal being affected by convergence errors, typically an equalizer in the steady state must not vary its coefcients by more than 1% within a symbol interval. Therefore another algorithm is needed to estimate '.
Adaptive method
The adaptive LMS algorithm is used for an instantaneous squared error dened as jek j2 D jak
D
xk
. pb/T . pb/
j .2 f 0 kT C'/ 2 O
(8.379)
The gradient of the function in (8.379) with respect to c. pb/ is equal to rc. pb/ jek j2 D D D where ek
. pb/ O D ek e j .2 f 0 kT C'k /
2ek .xk
. pb/
j .2 f 0 kT C'k / O
O 2ek e j .2 f 0 kT C'k / xk
. pb/
(8.380)
2ek
. pb/ . pb/ xk
(8.381)
C ek
. pb/ . pb/ xk
(8.382)
We now compute the gradient with respect to 'k . Let O D 2 f 0 kT C 'k O (8.383)
702
then jek j2 D .ak D .ak Therefore we obtain r' jek j2 D O @ . pb/ jek j2 D j yk e @
. pb/ j ek D D
yk /.ak yk
. pb/
D j
yk /
/.ak D
yk
. pb/ j
e /
(8.384)
ek j yk
. pb/ j
D 2Im[ek .yk
/ ]
(8.385)
D 2Im[ek yk ]
As ek D ak
(8.386)
We note that Im[ak D yk ] is related to the sine of the phase difference between ak D and yk , therefore the algorithm has reached convergence only if the phase of yk coincides (on average) with that of ak D . The law for updating the phase offset estimate is given by
' Im[ek yk ]
(8.387)
' Im[ak
D yk ]
(8.388)
The adaptation gain is typically normalized as ' D ' Q jak D j jyk j (8.389)
In general, ' is chosen larger than , so that the variations of ' are tracked by the carrier phase offset estimator, and not by the equalizer. In the ideal case, we have yk and yk D ak therefore the adaptive algorithm becomes: 'kC1 D 'k O O ' Im[e Q
j .' 'k / O De j .' 'k / O . pb/
D ak
De
j .2 f 0 kT C'/
(8.390)
(8.391)
'k / O
(8.392)
which is the equation of a rst-order phase-locked-loop (PLL) (see Section 14.7). For a constant phase offset, at convergence we get ' D '. O
703
8.14.2
As an example, we consider the previous QAM schemes in the case of transmission over a telephone channel, with passband in the range 3003400 Hz, for which a carrier f 0 in the range 17001800 Hz and a symbol rate 1=T of 2400 Baud are adopted. We note that for a transmit pulse h Tx of the square root raised cosine type with roll-off D 0:125 the transmission band goes from f 0 1350 Hz to f 0 C 1350 Hz. Using the conguration of Figure 8.4 with a sampling rate of 9600 Hz, equivalent to a sampling period Tc D T =4, the complex-valued scheme of a passband receiver, employing an FSE with sampling period equal to T =2 and a discrete-time phase-splitter, is illustrated in Figure 8.37. We note that, after the phase splitter lter, the frequency support of the received signal is halved, therefore we can sample with sampling period T =2. The loop lter updates the value of 'k according to (8.388). The voltage controlled O O oscillator (VCO) determines k D 2 f 0 kT C 'k , and computes the complex exponential e jk . The algorithm for coefcient updating is given by (8.382). Let us consider some simplications in the structure of Figure 8.37. From (8.360) let . pb/ . pb/ . pb/ g M;I and g M;Q , be, respectively, the in-phase and quadrature components of g M . Let c I c. pb/ .
. pb/
and c Q
. pb/
To determine yk , from the structure of Figure 8.37 we obtain the structure of Figure 8.38, where real-valued signals are considered. Overall we have six discrete-time convolutions. Combining the phase splitter and the equalizer of Figure 8.38, a single lter c. pb/ is used that operates with sampling period of the input signal equal to T =4, and sampling period of the output signal equal to T , as depicted in Figure 8.39. The LMS algorithm for coefcient adaptation is given by ckC1 D ck
. pb/ . pb/
C ek
. pb/
x4k
(8.393)
where c. pb/ is a complex-valued lter, and the passband input signal x is real-valued. The overall complexity of ltering and adaptation is lower as compared to the previous realization, although convergence is slower. A receiver structure with a baseband adaptive lter is depicted in Figure 8.40.
Figure 8.37. QAM passband receiver for transmission over telephone channels.
704
cI (pb) g (pb)
M,I
2 T 2
(pb) cQ
T + T y k,I
(pb) (pb)
r(t)
g (pb)
AA
x(pb) (t) T 4
xn
(pb) cQ
(pb) yk,Q
yk
g (pb)
M,Q
2 T 2 cI (pb) 2
T + + T
Figure 8.38. Implementation of the scheme of Figure 8.37 using real-valued signals and lters.
x(pb) (t)
xn T 4 c(pb) Q 4
T y(pb)
k
Figure 8.39. Efcient implementation of the receiver of Figure 8.37 combining phase splitter and equalizer.
cos 2 f0n
T 4
cI
T + T yk
r(t)
g (pb) AA
x(pb) (t) T 4
cQ xn cQ
T + + T
sin 2 f0n
T 4
cI
Figure 8.40. Efcient implementation of the QAM receiver using a baseband lter.
705
8.15
A fast method, in the sense that it requires a relatively short training interval, to determine the coefcients of a LE in the scheme of Figure 8.36 consists of using a periodic PN training sequence of period LT, equal to the time span of the equalizer lter delay line. In this case we speak of cyclic equalization, because the received signal, even though it is distorted, in the absence of noise is periodic of period LT [20]. We consider equalizers with N coefcients and sampling period of the input signal equal to T =F0 , with F0 > 1, because of the robustness they offer with respect to the choice of the timing phase. As it must be N .T =F0 / D L T , we have N D F0 L (8.394)
With reference to Figure 8.36 and (8.364), we consider the case in which after the phase . pb/ is shifted to baseband using a nominal carrier frequency f 0 . From splitter, the signal xn (8.364) and (8.357), we dene xn D xn N D
. pb/
T j2 f 0 n F
C1 X `D 1
a` h n
`F0 e
T j 2 1 f n F C1'
0
(8.395) C wn N
where 1 f and 1' represent, respectively, the carrier frequency offset and phase offset, both unknown quantities. In (8.395), the noise wn is given by N Q wn D wn e N
T j2 f 0 n F
0
(8.396)
where wn is dened in (8.366). Q The scheme of Figure 8.36 is redrawn as in Figure 8.41.
706
a` h n
`F0
(8.397)
at the instant the equalizer outputs the sample yk , the samples stored in the delay line of N the lter c are given by N x k F0 N
i
D vk F0 i e
T j 2 1 f .k F0 i / F C1' 0
C wk F0 N
i D 0; 1; : : : ; N
(8.398)
Choosing as training sequence fa` g the repetition of a PN sequence with period L, the sequence fvn g is periodic with period N D F0 L; in particular, vk F0
N i
D vk F0
(8.399)
Consequently, from (8.395), neglecting the noise wn , the sample that leaves the equalizer N N delay line, x k F0 N , differs from the sample that enters, x k F0 , by a phase rotation equal to N 2 1 f N .T =F0 / D 2 1 f L T . On the other hand, noise samples that are separate by an interval equal to LT are uncorrelated. We consider now two problems: detection of the beginning of the transmission and efcient computation of the equalizer coefcients.
n F0
(8.400)
ON where Nw is the number of observations utilized and M x is an estimate of the statistical power of xn . N d We underline that in (8.400), as 1 f is unknown, we use an estimate of the value 1 f , d. This choice yields which can be derived by minimizing M with respect to 1 f dopt 1f 1 arg D 2 L T
Nw 1 X nD0
! x k F0 N N n F0 x k F0 N n F0 (8.401)
1 d Because < arg < , j1 f opt j is smaller than 2L T ; this fact sets a limit on 1 f . For example, for a lter with L D 32 and 1=T D 2400 Baud it must be j1 f j<37:5 Hz. For dopt given by (8.400), the value of M is computed. If it falls below a certain 1f threshold then we detect the presence of a full period of the training signal in the equalizer lter delay line, and we trigger the procedure for computing the equalizer coefcients.
707
A statistical analysis leads to the following relations [21] 2 2 N E [ M j training signal present ] D w 1 Mx N and 2 2 N E [ M j training signal absent ] D w Mx N
1 2Nw s
(8.402)
2 Nw
! (8.403)
2 Considering that in the presence of the training signal we have Mx w , whereas in the N N 2 , the metric in (8.402) is much lower than that in absence of the training signal Mx ' w N N (8.403). In [21] it is observed that a value of Nw D 8 is adequate to reach a false alarm probability, i.e. the probability of detecting the training sequence in the presence of noise only, of 10 4 and a probability of missed detection, i.e. the probability of not detecting the training sequence, of 10 6 , even in the presence of severe distortion introduced by the telephone channel.
ci vk F0
D akmod L
8k
(8.404)
An estimate vn of the sample vn is obtained by rst compensating for the frequency O offset in xn , and successively averaging the received signal over Nv periods N vk F0 O
i
X 1 Nv 1 x k F0 N Nv mD0
j2 dopt k 1f
i mN e i F0 T
T j2 dopt .k F0 i m N / F 1f
(8.405) 1 Nv
Nv 1 X mD0
De
x k F0 N
i mN e
j2 dopt m L T 1f
i D 0; 1; : : : ; N
(8.406)
ci vk F0 O
D ak
k D 0; 1; : : : ; L
(8.407)
708
We now take the DFT of both members in the previous equation. Let Ap D
L 1 X kD0
ak e
kp j2 L
p D 0; 1; : : : ; L
(8.408)
Cq D
N 1 X i D0
ci e
iq j2 N
q D 0; 1; : : : ; N
(8.409)
O Vq D
N 1 X i D0
vi e O
iq j2 N
q D 0; 1; : : : ; N
(8.410)
The above relation represents a system of L linear equations and N D F0 L unknowns fCq g. Among the innite number of solutions we choose the one that minimizes the lter energy
N 1 X i D0
jci j2 D
X 1 N 1 jCq j2 N qD0
(8.412)
For the minimization of (8.412), with the constraint (8.411), the method of the Lagrange multipliers may be used. We consider here the particular case F0 D 2. For the general case we refer to [21]. For F0 D 2, we have N D 2L, and the solution is given by Cq D O 2Aqmod L Vq O O jVqmod L j2 C jV.LCq/mod L j2
iq X 1 N 1 Cq e j2 N N qD0
q D 0; 1; : : : ; N
(8.413)
Summarizing, the algorithm for the computation of the coefcients of an FSE with sampling period of the input signal equal to T =2 includes the following steps: 1. record Nv sequences of 2L samples from the equalizer delay line; 2. remove the phase rotation as indicated by (8.405); 3. perform a 2L-point DFT to obtain fVq g (fA p g may be pre-computed); 4. compute fCq g as indicated by (8.413); 5. perform a 2L-point inverse DFT to obtain fci g; 6. adjust the coefcients as indicated by (8.406).
709
to compensate for the phase introduced by the shift of the equalizer coefcients. After transition to data mode, the various parameters are updated by adaptive algorithms, similar to those derived in Section 8.14.1. In particular, to update the lter coefcients, from (8.382), we have N ckC1 D .1 N / ck C ek x N Nk (8.417)
where 0 < < 1 is a parameter that allows the LMS algorithm to track small variations of the channel characteristics (see (3.125)). To update the phase estimate, from (8.387) we have kC1 D k
' Im[ek yk ] C 2 T d 1f
(8.418)
d where 1 f , initialized with dopt given by (8.401), is updated by a second order PLL (see 1f Section 14.7).
710
8.16
LE and DFE in the frequency domain with data frames using cyclic prex
We consider now an equalization method, which is attractive for the low complexity that it requires for determining the lter coefcients and for performing the ltering operations. From the model of Figure 8.7, we obtain the discrete-time model of Figure 8.42, where the sampling period of the equalizer input signal is equal to T =2, and the polyphase representations of the sequences x, h, and c are employed; in particular we have the following relations. Overall impulse response at the receive lter output, with sampling period equal to T =2, fh i g h .`/ D h 2mC` m i D 0; 1; : : : ; 2Nh m D 0; 1; : : : ; N h 1 ` D 0; 1 (8.419)
Impulse response of the equalizer with sampling period of the input signal equal to T =2, fci g i D 0; 1; : : : ; N 1 N even (8.421)
Figure 8.42. (a) Discrete-time model; (b) overall model employing the polyphase representation.
711
` D 0; 1
(8.422)
Sequence of samples fxn g at the receive lter output with polyphase components .0/ .1/ fx k D x2k g and fx k D x2kC1 g. Q .0/ White noise sequence fwn g with PSD N0 , and polyphase components fwk g Q .1/ and fwk g. Q Let fak g, k D 0; 1; : : : ; M 1, be the transmitted sequence of symbols. To simplify the implementation of the equalizer, we consider the transmission of an extended sequence of . px/ symbols fak g, obtained by partially repeating fak g [23] ( k D 0; 1; : : : ; M 1 ak . px/ ak D (8.423) k D 1; : : : ; Npx aMCk In (8.423) Npx is the length of the cyclic prex, which is related to Nh , length of the channel impulse response in number of symbol periods T , so that it results Npx Nh 1 (8.424) We note that (8.423) assumes a transmission data frame such that, between blocks of data, there is a guard period within which the data are partially repeated. For a given bandwidth of the transmission channel, or rather for a given symbol rate, the system introduces an overhead of Npx symbols every M information symbols. .`/ From the received sequences fx k g, for k D Npx ; : : : ; 1; 0; 1; : : : ; M 1, ` D 0; 1, we omit the rst Npx samples, and we consider the sequence ( .`/ xk k D 0; 1; : : : ; M 1 .`/ (8.425) zk D 0 elsewhere We also introduce the following vectors with M components and the corresponding DFTs: a D [a0 ; : : : ; aM
1] T T
1]
1 ; 0; : : : ; 0] 1] T
.`/ .`/ H.`/ D [H0 ; : : : ; HM .`/ .`/ c.`/ D [c0 ; : : : ; c.N =2/
D DFT[h.`/ ]
T
1 ; 0; : : : ; 0]
Z .`/ D DFT[z.`/ ]
712
y D [y0 ; : : : ; yM Y D DFT[y]
1]
(8.434) (8.435)
It is easy to verify that if (8.424) holds, the same conditions as in (1.116) are veried, therefore we have the relation
.`/ .`/ Zm D Am Hm
m D 0; 1; : : : ; M
(8.436)
m D 0; 1; : : : ; M
(8.437)
with
.1/ N.1/ Cm D Cm e m j2 M
(8.438)
The receiver structure with a linear equalizer in the frequency domain is illustrated in Figure 8.43; the convolution between x and c is substituted by three M-point DFTs. The attractiveness of this structure resides in the simplicity of the determination of the equalizer coefcients to be used in (8.437). A rst method, described in Section 8.15, consists of adopting a training sequence of suitable length. Then from the DFTs of the various signals we determine the DFT of the sequence c (see (8.413)). As an alternative, proposed 1 in [24], we describe the MSE method (8.12) that for f D q 2MT =2 , q D 0; 1; : : : ; 2M 1, yields the DFT of the optimum FSE with sampling period of the input signal equal to T =2. In this case, we assume as known the impulse response fh i g and also its 2M-point DFT,
Figure 8.43. Structure of a linear equalizer in the frequency domain with data frames using cyclic prex.
713
1 f Dq 2MT =2
2 1 N0 C a 2 [jHq j2 C jHqCM j2 ]
q D 0; 1; : : : ; 2M
(8.440) 1, (8.441)
.Cm C CmCM /
and
.1/ Cm D 1 2 1 2
e j2 2M Cm C e j 2M .mCM/ CmCM
m
(8.442) (8.443)
e j M .Cm
CmCM /
e N.1/ Cm D
m j M
.Cm
CmCM /
(8.444)
We note that, exploiting the data frame structure with cyclic prex, the implementation of Figure 8.43 uses DFTs with a number of samples lower than that required by the general frequency domain method illustrated in Figure 3.22. A frequency domain DFE, which utilizes a known data sequence as guard interval, rather than a prex, has been proposed in [25]. In general, its performance is much better than that of the above LE conguration. In [25] design methods with a reduced complexity are also proposed for the direct design of the FF lter in the frequency domain.
8.17
Using Monte Carlo simulations (see Appendix 7.E) we give a comparison, in terms of Pbit as a function of the signal-to-noise 0, of the various equalization and data detection methods described in the previous sections. We refer to the system model of Figure 8.5 with an overall impulse response fh n g having ve coefcients, as given in Table 1.4 on page 26, 2 2 and AWGN noise w. Recalling the denition of the signal-to-noise 0 D a rh .0/=w , we Q Q examine four cases.
714
We anticipate that the ZF equalizer is designed by the DFT method (see Section 8.7). We also introduce the abbreviation DFE-VA to indicate the method consisting of the FF lter of a DFE, followed by MLSD implemented by the VA, with M M2 states determined by the impulse response n D nCD for n D 0; : : : ; M2 : this technique is commonly used to shorten the overall channel impulse response and thus simplify the VA (see case (2a) on page 679). 1. ZF with N D 7 and D D 0; 2. LE with N D 7 and D D 0; 3. DFE with M1 D 7, M2 D 4, and D D 6; 4. VA with 44 D 256 states and path memory depth equal to 15, i.e. approximately three times the length of fh n g; 5. DFSE with J1 D J2 D 4, J3 D J4 D 1 (16 states); 6. DFE-VA with M1 D 7, M2 D 2, and D D 0 (VA with 4 M2 D 16 states). The performance of the various receivers is illustrated in Figure 8.44; for comparison, the performance achieved by transmission over an ideal AWGN channel is also given.
10
10
10
Pbit
10
4
10
10
10
11 (dB)
12
13
14
15
16
Figure 8.44. Bit error probability, Pbit , as a function of 0 for QPSK transmission over a minimum phase channel, using various equalization and data detection methods.
715
10
10
Pbit
10
4
10
10
10
11 (dB)
12
13
14
15
16
Figure 8.45. Bit error probability, Pbit , as a function of 0 for QPSK transmission over a non-minimum phase channel, using various equalization and data detection methods.
716
10
10
Pbit
10
4
10
10
11
12
13
14
15
16
17
18
19
20
(dB)
Figure 8.46. Bit probability error, Pbit , as a function of 0 for 8-PSK transmission over a minimum phase channel, using various equalization and data detection methods.
717
10
10
10
Pbit
10
4
10
10
11
12
13
14
15
(dB)
16
17
18
19
20
Figure 8.47. Bit probability error rate, Pbit , as a function of 0 for 8-PSK transmission over a non-minimum phase channel, using various equalization and data detection methods.
In these simulations, the error probability for a DFE-VA is in the range between 0.08 and 0.2 and it is not shown; the performance of the various receivers is illustrated in Figure 8.47. By comparison of the results of Figure 8.46 and Figure 8.47 we observe that the RSSE and DFSE may be regarded as valid approximations of the VA as long as the overall channel impulse response is minimum phase.
8.18
The various equalization and data detection methods described in this chapter can be extended to systems in which the transmission of the information message takes place over several channels. The task of the receiver is to suitably combine the various received signals and to form a signal at the decision point with a better signal-to-noise ratio, or bit error probability, as compared to the case of transmission over a single channel. These structures are widely used in radio systems that are subject to channel fading (see Section 4.6). Diversity is realized by using more antennas at the receiver and/or at the transmitter. The more uncorrelated are the channels, the higher is the diversity gain. In indoor and urban environments this requires a spacing between antennas of at least =2, whereas in rural environments the spacing may be a multiple of . In spread spectrum systems (see Chapter 10) the diversity is generated by the multipath channel itself and we speak of multipath diversity. In general, using suitable techniques, at the receiver it is possible to compose a signal in which the effect of fading is reduced.
718
Figure 8.48. Receiver with two receive antennas for at fading radio channels.
Antenna arrays
Let us consider the scheme of Figure 8.4. We generalize it to the case of two receive antennas: thus, we obtain the scheme of Figure 8.48, where the receive lter for each branch is the matched lter to the received pulse. For a non-dispersive transmission channel, assuming absence of ISI, for a single transmit antenna and N ARc receive antennas, at the i-th antenna branch the sampled signal, with sampling period equal to T , is given by
.i .i / x k / D A ak gC;0 C wk / C vk / Q .i Q .i
(8.445)
where ak is the desired symbol, A is the amplitude of the desired pulse at the output of the downsampler (in the scheme of Figure 8.48 A D E h , the energy of the trans.i / mitted pulse h Tx ), gC;0 is a complex coefcient representing the at fading channel with
2 samples, having variance w.i/ D N0 E h . The sample vk / represents the interference on the Q .i Q i-th branch due to undesired signals using the same carrier as the desired signal. If N I is the number of interfering signals, that are assumed synchronous with the desired signal, we have .i / E[jgC;0 j2 ] D 1, and fwk / g, i D 1; : : : ; N ARc , are uncorrelated sequences of i.i.d. noise Q .i
vk / D Q .i
NI X jD1
A. j/ ak gC;0
. j/
. j/
.i; j/
(8.446)
With regard to the j-th interfering signal, ak is the generic symbol of the transmitted message, A. j/ is the amplitude of the interfering pulse at the output of the downsampler, .i; j/ and gC;0 represents the at fading channel between the j-th interfering antenna and the i-th receiving antenna. We also assume that there is no Doppler spread. Therefore the channels are time invariant for the duration of the transmission. Finally, we recall the reference signal-to-noise ratio at the decision point for an ideal AWGN channel, MF D 2E0h , given by (7.113). N
719
The signal at the decision point is usually a linear combination of samples given by (8.445), that is
N ARc
yk D
X
i D1
c.i / x k
.i /
(8.447)
where fc.i / g, i D 1; : : : ; N ARc , are suitable coefcients. In general, however, the combination .i can be made either directly on the received signals frC / .t/g, i D 1; : : : ; N ARc , thus realizing a pre-detection combiner, or on signals after the matched lter as in (8.447), realizing in this case a post-detection combiner. We emphasize that the scheme of Figure 8.48 represents the simplest case, with only two antennas, of a structure, or array, that in general has N ARc antennas. Depending on the placement of elements, arrays are classied as: 1) linear if the elements are aligned on a straight line; 2) circular if the elements are placed on a circle; 3) planar if the elements are placed on a grid. More complex structures also exist.
Combining techniques
For the selection of the coefcients in (8.447) we consider two switching techniques and three combining techniques. To simplify the analysis we assume vk / absent or included Q .i in wk / . Q .i 1. Selective combining. Only one of the received signals is selected. Let i SC be the branch corresponding to the received signal with highest statistical power, i SC D arg
.i/ i 2f1;:::;N ARc g rC
max
(8.448)
where the different powers are estimated using a training sequence. In some cases, in place of the statistical power the bit error probability or the receive signal-to-noise ratio8 is used as the decision parameter. Based on the decision (8.448), the receiver selects the antenna .i i SC and consequently extracts the signal x k SC / aligned in phase. With reference to (8.447), this method is equivalent to selecting ( .i / gC;0 i D i SC .i / (8.449) c D 0 i 6D i SC and
SC yk D gC;0 x k SC .iSC / 2 .i SC / .i / D gC;0 A ak C gC;0 wk SC Q
.i
.i
(8.450)
MF
.iSC / 2 g
C;0
(8.451)
This parameter can be estimated by the estimate of the channel impulse response (see Appendix 3.B)
720
A variant of this technique consists in selecting only two or three signals; next, their combination takes place. 2. Switched combining. Another antenna is selected only when the statistical power of the signal, or equivalently the bit error probability, or the receive signal-to-noise ratio, drops below a given threshold. Once a new antenna is selected, the signal is processed as in the previous case. 3. Equal gain combining (EGC). we have c.i / D e It results in In this case the signals are only aligned in phase; therefore
.i/
j arg.gC;0 /
D e j arg.gC;0 /
.i/
i D 1; : : : ; N ARc
(8.452)
(8.453)
which yields 12 X .i / g A @ 0
N ARc i D1 MF C;0 EGC
N ARc
(8.454)
This technique is often used in receivers of DPSK signals (see Section 6.5); in this case the combining is obtained by summing the differentially demodulated signals on the various branches. 4. Maximal ratio combining (MRC). We assume absence of interferers. The MRC criterion consists in maximizing the signal-to-noise ratio at the decision point. Substituting (8.445) in (8.447) and taking the expectation with respect to the message and the various noise signals, we get 2 N ARc X .i / .i / c gC;0 i D1 2 (8.455) D 2E h N ARc X 2 c.i / 2 .i/ w Q
i D1
Using the Schwarz inequality (see page 4), the signal-to-noise in (8.455) is maximized for c.i / D K where K is a constant.
.i / gC;0
w.i/ Q
i D 1; : : : ; N ARc
(8.456)
721
Because the noise signals of the various branches have the same variance, the choice
.i / c.i / D K gC;0
(8.457)
, given by 0
MRC
MF
1 X .i / 2 g A @
N ARc i D1 C;0
(8.458)
Introducing the signal-to-noise ratio of the i-th branch .i / 2 .i / D MF gC;0 (8.458) can be written as
N ARc MRC
(8.459)
X
i D1
.i /
(8.460)
that is, the total signal-to-noise ratio is the sum of the signal-to-noise ratios of the individual branches. It is interesting to note that the choice (8.457) is also the solution of the maximum likelihood criterion associated with signals (8.445). In the case of uncorrelated channels with a Rayleigh statistic, the performance, in terms of bit error probability of the various combining techniques can be obtained analytically by applying the technique described in Section 6.12 [26, 27, 28]. 5. Optimum combining (OC). The MRC criterion performs well in situations where the desired signal power is much greater than the power of the interfering signals; otherwise the effect of interference must be considered. In the OC criterion the ratio between the power of the desired signal and the power of the interference plus noise (SINR) is maximized. In practice the coefcients fc.i / g, i D 1; : : : ; N ARc , are determined by the Wiener formulation, where yk is the sample at the decision point and ak is the desired symbol. Dening
.1/ xk D [x k ; : : : ; x k .N ARc / T
(8.461)
(8.462)
(8.463)
722
(8.465)
In many applications, starting from estimates of R and p obtained by equations similar to (2.130) and (2.131) (see also the covariance method of page 148), the estimate of R 1 is 3 obtained directly with a computational complexity O.N ARc /. This method is termed direct matrix inversion (DMI). As an alternative, or next to a rst solution through the DMI method, the LMS or RLS iterative methods may be used (see Chapter 3), that present the advantage of tracking the system variations. A graphic method that illustrates the effectiveness of the MRC and OC techniques is the array radiation diagram [29], which shows a main lobe in the direction of the desired user and zeros along the direction of the interferers: in fact a receiver with N ARc antennas and N I interferers has the same performance of a receiver with .N ARc N I / antennas and no interferers [30, 31]. However, if the interferers are known, cancellation methods give better performance.
(8.466)
(8.467)
the solution is of the type (8.465). Also in this case a DFE structure, or better an MLSD receiver, yields improved performance at the cost of a greater complexity. For an in-depth study we refer to the bibliography [32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49]. We point out that the optimum structure of a combiner/linear equalizer is given by a bank of lters, each matched to the overall receive pulse at the antenna output, whose outputs are then summed; the signal thus obtained is equalized by an LE or a DFE [50]. We also observe that in mobile radio systems, the use of multiple antennas generally occurs at the base-station, as the mobile is limited both in power consumption and dimension.
Diversity in transmission
We now discuss some diversity methods in which N ATx transmit antennas and one receive antenna are employed [51, 52, 53].
723
First, we distinguish close-loop methods, with feedback from receiver, from open-loop methods, without feedback. Obviously in the rst case there is a return channel to communicate the selected parameters to the transmitter; the drawback of these methods is that, besides requiring a certain capacity, the parameters of the transmission channel are known with a certain delay. 1. Selection transmission diversity. As illustrated in Figure 8.49, the idea is to select in transmission the antenna that yields the best value of the received power, or bit error rate, or signal-to-noise ratio. To identify this antenna, a training sequence must be transmitted by each antenna; if the training sequences are orthogonal, they can be simultaneously transmitted. The receiver then communicates to the transmitter the antenna to be employed in the next slot or frame (see Section 6.13.2). 2. Transmit array. The signals sent over the various antennas are multiplied by coefcients fc.i / g, i D 1; : : : ; N ATx , computed at the receiver (see Figure 8.50). At the receiver the criterion can be the MRC or the OC. In order not to increase the average transmitted power, the coefcients must be scaled so that
N ATx i D1
X 2 c.i / D 1
(8.468)
In [52] a method is presented to determine the coefcients which minimize the receiver bit error probability under a constraint on the total transmitted power or on the power transmitted by each antenna.
Tx,1 Rc ak s(t) Tx,2 ^ ak
hTx
DEMOD
DEMOD
724
Rc
* hTx,1 (-t)
x (1)
k
c (1) yk ^ ak
x (2)
k
c (2)
3. Orthogonal transmission diversity. The symbol ak is modulated with two orthogonal pulses, h Tx;1 .t/ and h Tx;2 .t/, as illustrated in Figure 8.51. The receiver employs two matched lters, that output the signals
.i .i / x k / D A ak gC;0 C wk / Q .i
i D 1; 2
(8.469)
(8.470)
we get
.1/ .2/ yk D c.1/ x k C c.2/ x k .1/ .2/ .1/ .1/ .2/ .2/ D A ak .jgC;0 j2 C jgC;0 j2 / C gC;0 wk C gC;0 wk Q Q
(8.471)
If the noise signals wk / , i D 1; : : : ; N ATx , are uncorrelated, and if each antenna can Q .i transmit a signal with maximum power, this scheme has the same performance of an MRC scheme, in terms of ; for the same total transmitted power, instead, it loses 10 log10 N ATx dB. Another drawback lies in the use of at least two orthogonal signals per user. A variant of the system is proposed in [51]. 4. Delay diversity. Let s.t/ be the modulated signal, as shown in Figure 8.52. The i-th antenna transmits the delayed signal s.t iT/ i D 0; : : : ; N ATx 1 (8.472)
At the receiver, the message is detected from the resulting signal with ISI by MLSD [54]. 5. Time switched transmit diversity. As illustrated in Figure 8.53, the transmitter selects in turn the antenna on which to transmit a symbol sequence. The method is much simpler than the previous one, however, it requires the use of channel coding and interleaving (see Chapter 11) to recover the errors introduced by some channels.
725
hTx
Tx,1
Rc ^ ak
DEMOD Tx,NA
Tx
Rc ^ ak
DEMOD
hTx
6. Space-time transmit diversity. The basic scheme is illustrated in Figure 8.54 [53]. The message fak g is transmitted over two antennas: over antenna 1 the transmitted signal is modulated by the data sequence a2n a2nC1
a2nC1 ; : : :
(8.473) (8.474)
726
For a receiver with matched lter to the pulse h Tx , at the decision point we have the signal 8 < A.a2n g .1/ C a2nC1 g .2/ / C w2n Q for k D 2n C;0 C;0 xk D (8.475) .1/ : A. a .2/ Q for k D 2n C 1 2nC1 gC;0 C a2n gC;0 / C w2nC1 Assuming the channels are known, we consider the combination of the samples
.1/ .2/ y2n D gC;0 x2n C gC;0 x2nC1
(8.476)
and
.2/ y2nC1 D gC;0 x2n .1/ gC;0 x2nC1
(8.477)
It is easy to verify that 8 < a2n .jg .1/ j2 C jg .2/ j2 / C g .1/ w2n C g .2/ w C;0 C;0 C;0 Q C;0 Q 2nC1 yk D .1/ 2 .2/ 2 .2/ .1/ :a Q gC;0 w2nC1 Q 2nC1 .jgC;0 j C jgC;0 j / C gC;0 w2n
A comparison with the MRC technique leads to the same considerations made on (8.471). We conclude this section by mentioning diversity techniques with more than one transmit and receive antenna, called space-time coding techniques, whereby the message is coded by using suitable channel codes [55, 56, 57].
Bibliography
[1] G. Ungerboeck, Fractionally tap-spacing and consequences for clock recovery in data modems, IEEE Trans. on Communications, Aug. 1976. [2] R. D. Gitlin and S. B. Weinstein, Fractionally-spaced equalization: an improved digital transversal equalizer, Bell System Technical Journal, vol. 60, Feb. 1981. [3] J. E. Mazo and J. Salz, Probability of error quadratic detector, Bell System Technical Journal, vol. 44, Nov. 1965. [4] J. G. Proakis, Digital communications. New York: McGraw-Hill, 3rd ed., 1995. [5] L. R. Rabiner, A tutorial on hidden Markov models and selected applications in speech recognition, IEEE Proceedings, vol. 77, pp. 257285, Feb. 1989. [6] L. R. Bahl, J. Cocke, F. Jelinek, and J. Raviv, Optimal decoding of linear codes for minimizing symbol error rate, IEEE Trans. on Information Theory, vol. 20, pp. 284 287, Mar. 1974.
8. Bibliography
727
[7] A. J. Viterbi, An intuitive justication and simplied implementation of the MAP decoder for convolutional codes, IEEE Journal on Selected Areas in Communications, vol. 16, pp. 260264, Feb. 1998. [8] G. Ungerboeck, Adaptive maximum likelihood receiver for carrier modulated data transmission systems, IEEE Trans. on Communications, vol. 22, pp. 624635, May 1974. [9] G. E. Bottomley and S. Chennakeshu, Unication of MLSE receivers and extension to time-varying channels, IEEE Trans. on Communications, vol. 46, pp. 464472, Apr. 1998. [10] D. G. Messerschmitt and E. A. Lee, Digital communication. Boston, MA: Kluwer Academic Publishers, 2nd ed., 1994. [11] K. M. Chugg and A. Anastasopoulos, On symbol error probability bounds for ISI-like channels, IEEE Trans. on Communications, vol. 49, pp. 17041709, Oct. 2001. [12] W. U. Lee and F. S. Hill, A maximum-likelihood sequence estimator with decision feedback equalization, IEEE Trans. on Communications, vol. 25, pp. 971979, Sept. 1977. [13] G. J. Foschini, A reduced-state variant of maximum-likelihood sequence detection attaining optimum performance for high signal-to-noise ratio performance, IEEE Trans. on Information Theory, vol. 24, pp. 505509, Sept. 1977. [14] J. B. Anderson and S. Mohan, Sequential coding algorithms: a survey and cost analysis, IEEE Trans. on Communications, vol. 32, pp. 169176, Feb. 1984. [15] M. V. Eyuboglu and S. U. H. Qureshi, Reduced-state sequence estimator with set partitioning and decision feedback, IEEE Trans. on Communications, vol. 36, pp. 13 20, Jan. 1988. [16] M. V. Eyuboglu, Reduced-state sequence estimator for coded modulation on intersymbol interference channels, IEEE Trans. on Communications, vol. 7, pp. 989995, Aug. 1989. [17] R. E. Kamel and Y. Bar-Ness, Reduced-complexity sequence estimation using state partitioning, IEEE Trans. on Communications, vol. 44, pp. 10571063, Sept. 1996. [18] W. Sheen and G. L. Stuber, Error probability for reduced-state sequence estimation, IEEE Journal on Selected Areas in Communications, vol. 10, pp. 571578, April 1992. [19] R. Raheli, A. Polydoros, and C.-K. Tzou, Per-survivor processing: a general approach to MLSE in uncertain environments, IEEE Trans. on Communications, vol. 43, pp. 354364, Feb.Apr. 1995. [20] K. H. M ller and D. A. Spaulding, Cyclic equalization: a new rapidly converging u equalization technique for synchronous data communication, Bell System Technical Journal, vol. 54, pp. 369406, Feb. 1975.
728
[21] P. R. Chevillat, D. Maiwald, and G. Ungerboeck, Rapid training of a voiceband data-modem receiver employing an equalizer with fractional-T spaced coefcients, IEEE Trans. on Communications, vol. COM35, pp. 869876, Sept. 1987. [22] G. Ungerboeck, D. Maiwald, H. P. Kaeser, and P. R. Chevillat, The SP16 signal processor, in Proc. Int. Conf. on Acoustics, Speech and Signal Processing, pp. 16.2.1 16.2.4, Mar. 1984. [23] H. Sari, G. Karam, and I. Jeanclaude, An analysis of orthogonal frequency-division multiplexing for mobile radio applications, in Proc. 1994 IEEE Vehicular Technology Conference, vol. 3, Stockholm, Sweden, pp. 16351639, June 810 1994. [24] M. Huemer, L. Reindl, A. Springer, and R. Weigel, Frequency domain equalization of linear polyphase channels, in Proc. 2000 IEEE Vehicular Technology Conference, Tokyo, Japan, May 2000. [25] N. Benvenuto and S. Tomasin, On the comparison between OFDM and single carrier modulation with a DFE using a frequency domain feedforward lter, IEEE Trans. on Communications, 2002. [26] T. S. Rappaport, Wireless communications: principles and practice. Englewood Cliffs, NJ: Prentice-Hall, 1996. [27] G. L. Stuber, Principles of mobile communication. Norwell, MA: Kluwer Academic Publishers, 1996. [28] T. Eng, N. Kong, and L. B. Milstein, Comparison of diversity combining techniques for Rayleigh-fading channels, IEEE Trans. on Communications, vol. 44, pp. 1117 1129, Sept. 1996. [29] C. G. Someda, Electromagnetic waves. London: Chapman & Hall, 1998. [30] J. H. Winters, Optimum combining in digital mobile radio with cochannel interference, IEEE Journal on Selected Areas in Communications, vol. 2, pp. 528539, July 1984. [31] J. H. Winters, J. Salz, and R. D. Gitlin, The impact of antenna diversity on the capacity of wireless communication systems, IEEE Trans. on Communications, vol. 42, pp. 17401751, Feb.Apr. 1994. [32] W. C. Jakes, Microwave mobile communications. New York: IEEE Press, 1993. [33] B. Glance and L. J. Greenstein, Frequency-selective fading in digital mobile radio with diversity combining, IEEE Trans. on Communications, vol. 31, pp. 10851094, Sept. 1983. [34] W. K. Kennedy, L. J. Greenstein, and M. V. Clark, Optimum linear diversity receivers for mobile communications, IEEE Trans. on Vehicular Technology, vol. 43, pp. 4756, Feb. 1994.
8. Bibliography
729
[35] M. V. Clark, L. J. Greenstein, W. K. Kennedy, and M. Sha, MMSE diversity combining for wide-band digital cellular radio, IEEE Trans. on Communications, vol. 40, pp. 11281135, June 1992. [36] R. T. Compton, Adaptive antennas: concepts and performance. Englewood Cliffs, NJ: Prentice-Hall, 1988. [37] D. D. Falconer, M. Abdulrahman, N. W. K. Lo, B. R. Petersen, and A. U. H. Sheikh, Advances in equalization and diversity for portable wireless systems, Digital Signal Processing, vol. 3, pp. 148162, 1993. [38] G. J. Foschini, Layered space-time architecture for wireless communication in a fading environment when using multi-element antennas, Bell System Technical Journal, vol. 1, pp. 4159, Autumn 1996. [39] L. C. Godara, Application of antenna arrays to mobile communicationsPart I: Performance improvement, feasibility and system considerations, IEEE Proceedings, vol. 85, pp. 10311060, July 1997. [40] L. C. Godara, Application of antenna arrays to mobile communicationsPart II: Beam-forming and direction-of-arrival considerations, IEEE Proceedings, vol. 85, pp. 11951245, Aug. 1997. [41] I. P. Kirsteins and D. W. Tufts, Adaptive detection using low rank approximation to a data matrix, IEEE Trans. on Aerospace and Electronic Systems, vol. 30, pp. 5567, Jan. 1994. [42] R. Kohno, H. Imai, M. Hatori, and S. Pasupathy, Combination of an adaptive array antenna and a canceller of interference for direct-sequence spread-spectrum multipleaccess system, IEEE Journal on Selected Areas in Communications, vol. 8, pp. 675 682, May 1990. [43] W. C. Y. Lee, Effects on correlation between two mobile radio base-station antennas, IEEE Trans. on Communications, vol. 21, pp. 12141224, Nov. 1973. [44] R. A. Monzingo and T. W. Miller, Introduction to adaptive arrays. New York: John Wiley & Sons, 1980. [45] A. J. Paulraj and C. B. Papadias, Space-time processing for wireless communications, IEEE Signal Processing Magazine, vol. 14, pp. 4983, Nov. 1997. [46] J. Salz and J. H. Winters, Effect of fading correlation on adaptive arrays in digital wireless communications, IEEE Trans. on Vehicular Technology, vol. 43, pp. 1049 1057, Nov. 1994. [47] G. Tsoulos, M. Beach, and J. McGeehan, Wireless personal communications for the 21st century: European technological advances in adaptive antennas, IEEE Communications Magazine, vol. 35, pp. 102109, Sept. 1997.
730
[48] B. Widrow, P. E. Mantey, L. J. Grifths, and B. B. Goode, Adaptive antenna systems, IEEE Proceedings, vol. 55, pp. 21432159, Dec. 1967. [49] J. H. Winters, Smart antennas for wireless systems, IEEE Personal Communications Magazine, vol. 5, pp. 2327, Feb. 1998. [50] P. Balaban and J. Salz, Dual diversity combining and equalization in digital cellular mobile radio, IEEE Trans. on Vehicular Technology, vol. 40, pp. 342354, May 1991. [51] K. Rohani, M. Harrison, and K. Kuchi, A comparison of base station transmit diversity methods for third generation cellular standards, in Proc. 1999 IEEE Vehicular Technology Conference, Houston, USA, pp. 351355, May 1620 1999. [52] J. K. Cavers, Optimized use of diversity modes in transmitter diversity systems, in Proc. 1999 IEEE Vehicular Technology Conference, Houston, USA, pp. 17681773, May 1620 1999. [53] S. A. Alamouti, A simple transmit diversity technique for wireless communications, IEEE Journal on Selected Areas in Communications, vol. 16, pp. 14511458, Oct. 1998. [54] J. H. Winters, The diversity gain of transmit diversity in wireless systems with Rayleigh fading, IEEE Trans. on Vehicular Technology, vol. 47, pp. 119123, Feb. 1998. [55] V. Tarokh, N. Seshadri, and A. R. Calderbank, Space-time codes for high data rate wireless communication: performance criterion and code construction, IEEE Trans. on Information Theory, vol. 44, pp. 744765, Mar. 1998. [56] A. Naguib, V. Tarokh, N. Seshadri, and A. R. Calderbank, A space-time coding modem for high data rate wireless communications, IEEE Journal on Selected Areas in Communications, vol. 16, pp. 14591478, Oct. 1998. [57] V. Tarokh, A. Naguib, N. Seshadri, and A. R. Calderbank, Combined array processing and space-time coding, IEEE Trans. on Information Theory, vol. 45, pp. 11211128, May 1999. [58] L. E. Franks, Signal theory. Stroudsburg, PA: Dowden and Culver, revised ed., 1981. [59] S. Benedetto and E. Biglieri, Principles of digital transmission with wireless applications. New York: Kluwer Academic Publishers, 1999. [60] N. Al-Dhahir and J. M. Ciof, MMSE decision-feedback equalizers: nite-length results, IEEE Trans. on Information Theory, vol. 41, pp. 961975, July 1995.
731
Appendix 8.A
The rst part of this appendix introduces the calculus of variations [58]. This technique will be applied to solve two optimization problems as described in the second part of the appendix.
8.A.1
Calculus of variations
The minimization of a functional9 with respect to a continuous-time signal differs from the classical problem of the Wiener lter because the signal cannot be represented by a nite number of coefcients. In this section, using the denition of a complete basis for continuous-time signals, we introduce the gradient of linear and quadratic functionals.
Linear functional
Denition 8.2 Let x.t/, t 2 <, be a real-valued signal. We dene linear functional of x the following integral: Z C1 g.t/x.t/ dt (8.479) Lx D
1
where g.t/ is also a real-valued signal. By (1.20) we represent the signal x by referring to a complete orthornormal basis f i .t/g, i 2 I, with I nite or numerable set, that is X xi i .t/ (8.480) x.t/ D
i 2I
where xi D hx.t/;
i .t/i
Z D
C1 1
x.t/ i .t/ dt
i 2I
(8.481)
are the signal components with respect to the basis. Substitution of (8.480) in (8.479) yields " Z C1 X g.t/ xi Lx D
1
#
i .t/
dt (8.482)
X
i 2I
Z xi
i 2I C1 1
g.t/ i .t/ dt
732
where we assume the order between the summation and integration can be exchanged. Let Z C1 g.t/ i .t/ dt i 2I (8.483) gi D hg.t/; i .t/i D
1
If we introduce the vectors associated with the signals x.t/t 2 < g.t/t 2 < (8.482) becomes Lx D xT g Observing (2.24), the gradient of (8.486) is rx Lx D g whose i-th component is equal to [rx Lx ]i D gi We now introduce the gradient as a continuous-time signal, X rx Lx D [rx Lx ]i i .t/
i 2I
! x D [: : : ; x ! g D [: : : ; g
1 ; x0 ; x1 ; : : : ]
T T
(8.484) (8.485)
1 ; g0 ; g1 ; : : : ]
(8.486)
(8.487)
(8.488)
X
i 2I
gi
i .t/
(8.489)
D g.t/ using (8.485). As we will see later, it is convenient to express the functional in the frequency domain; applying the Parseval theorem, we get Z C1 Z C1 Lx D g.t/x.t/ dt D G. f /X . f / d f D LX (8.490)
1 1
Quadratic functional
Denition 8.3 Let x.t/, t 2 <, be a real-valued signal. We dene quadratic functional of x the integral Z C1 Z C1 Qx D x.t/A.t; /x. / d dt (8.492)
1 1
733
As in the previous case, if we represent the signal x by a complete orthornormal basis we obtain Qx D xT Cx where C is the matrix with coefcients Z C1 Z C1 ci; j D i .t/A.t; /
1 1
(8.493)
j . / d
dt
i; j 2 I
(8.494)
Dening Ai . / D hA.t; /;
i .t/i
Z D
C1 1
A.t; / i .t/ dt
j 2I
(8.495)
(8.496)
D hA.t; /;
j . /i
Z D
C1 1
A.t; /
j . / d
j 2I
(8.497)
A0j .t/
j . /
(8.498)
The gradient of (8.493) is not equal to that computed in (2.30): in fact, in general, the matrix C is not symmetric and the vector x is real-valued. Hence, we obtain rx Qx D Cx C .xT C/T D Cx C CT x D .C C CT /x In particular the i-th component of the gradient is given by X .ci; j C c j;i /x j [rx Qx ]i D
j2I
(8.499)
(8.500)
2 3 X X 4 .ci; j C c j;i /x j 5 D
i 2I j2I
(8.501)
i .t/
734
j . / C
j ./A.; / i . / d
d]x j
i .t/
X X Z
i 2I j2I
C1 Z C1 1 1
i ./[A.; /
C A.; /] d
j . / d
xj
i .t/
(8.502) Substituting (8.495) and (8.498) in the above equation, changing variables, and using (8.496) and (8.498), we obtain X X Z C1 0 rx Qx D [Ai . / C Ai . /] j . / d x j i .t/
i 2I j2I C1 1 1
Z D Z D
" X
i 2I
#2 X 0 [Ai . / C Ai . /] i .t/ 4 xj
j2I
3
j . /5 d
(8.503)
C1 1
Also in this case we express the quadratic functional in the frequency domain by applying the Parseval theorem: Z C1 Z C1 X . f /B. f; /X ./ d f d D QX (8.504) Qx D
1 1
where B. f; /, called frequency operator, is the two-dimensional Fourier transform of A.t; /, Z C1 Z C1 A.t; /e j2 f t eC j2 dt d (8.505) B. f; / D
1 1
From (8.503) the gradient of the quadratic cost function becomes Z C1 [B. f; / C B .; f /]X ./ d rX QX D
1
(8.506)
To conclude, we give an example of a quadratic functional that will be useful in the next subsection. Example 8.A.1 We consider the following quadratic functional in the frequency domain: QX D Z
C1 1
P. f /jX . f /j2 d f
(8.507)
735
1 [P. f /],
it can be veried that the corresponding time and frequency A.t; / D r.t / / (8.508) (8.509)
B. f; / D P. f /. f
Therefore from (8.506) the gradient of the functional (8.507) is given by rX QX D 2P. f /X . f / (8.510)
8.A.2
Receiver optimization
Using the results of the previous section, we are now able to prove equation (8.11). Referring to the scheme of Figure 7.12 reproduced in Figure 8.55 with qC .t/ D h T x gC .t/, we need to solve the following problem. Given HT x . f / and GC . f /, that is QC . f / D HT x . f /GC . f / we desire to determine G Rc . f / that minimizes the cost function (8.10) J D E[jyk The following assumptions are made. 1. The sequence fak g is WSS, with zero mean and autocorrelation ra .n/ D E[ak ak n ]. 2. The noise wC .t/ is WSS with zero mean and spectral density PwC . f /. 3. fan g and wC .t/ are statistically independent. From (7.65), the signal at the sampler output is given by yk D where, from (7.67), h i D q R .t0 C i T /
qC ak rC (t) hTx gC wC (t)
Figure 8.55. Baseband transmission system.
C1 X iD 1
(8.511)
ak j2 ]
(8.512)
ai h k
C w R;k
(8.513)
(8.514)
t0 +kT gRc
yk
^ ak =Q[yk ]
736
with Fourier transform F[h i ] D using the relations H. f / D Q R . f /e j2 f t0 and, from (7.48), Q R . f / D QC . f /G Rc . f / Moreover, the variance of fw R;k g is given by (7.75). Substitution of (8.513) in the expression of J yields 2 2 3 C1 X ai h k i C w R;k ak 5 J D E 4 i D 1 D
C1 X C1 X C1 X iD 1
hi e
j2 f i T
C1 1 X H f T `D 1
` T
(8.515)
(8.516)
(8.517)
E[ai a h k i h j k
2 j ] C w R
2 C a C 2
C1 X i D1
E[ai h k i w ] R;k
(8.518)
i D 1 jD 1
C1 X iD 1
E[ai h k i ak ]
2E[w R;k ak ]
From the above assumptions, the fourth and the sixth term in the last expression are identically zero; therefore
2 J D w R C C1 X C1 X
ra .i
j/h k i h k
C1 X iD 1
ra .i
k/h k
2 C a
(8.519)
i D 1 jD 1
i and q D k
p/h p h q
ra .q ra . p
pD 1 qD 1 2 D w R C C1 X C1 X
q/h h q p
C1 X pD 1
(8.520)
2 ra . p/h C a p
pD 1 qD 1
It is convenient to express the above terms in the frequency domain, using the following relations: power spectral density of the message Pa . f / D T
C1 X nD 1
ra .n/e
j2 f nT
(8.521)
737
(8.522)
where H. f / D F[h.t/]. Now we rewrite the three terms on the right-hand side of (8.520). 1. Recalling that the noise is ltered only by the receive lter, we have Z C1 Z C1 jH. f /j2 2 w R D PwC . f /jG Rc . f /j2 d f D PwC . f / df jQC . f /j2 1 1 2. From (8.521) and (8.522), using (8.515), it turns out
C1 X C1 X
(8.523)
ra . p
C1 X
q/h h q p ra . p
C1 X qD 1
pD 1 qD 1
C1 X
Z q/ "
C1 1
H . f /e
j2 f pT
d f hq # hq d f
pD 1 qD 1
Z D
C1 1
H .f/
C1 X pD 1
ra . p
q/e
j2 f pT
(8.524)
1 T
Z Z
C1 1
H . f /
C1 X qD 1
Pa . f /e
C1 X `D 1
j2 f qT
hq d f ` T df
1 D 2 T
C1 1
H . f /Pa . f /
ra . p/h D p D
C1 X
ra . p/
C1 1
H . f /e
j2 pT
df (8.525)
pD 1 Z 2 C1
H . f /Pa . f / d f
Substitution of (8.523), (8.524), and (8.525) in (8.520) yields Z C1 jH. f /j2 PwC . f / df J D jQC . f /j2 1 Z C1 C1 X ` 1 df H . f /Pa . f / H f C 2 T T 1 `D 1 Z 2 C1 H . f /Pa . f / d f T 1
2 C a
(8.526)
738
a) The rst term is a quadratic cost function. From (8.510) the gradient with respect to H. f / is given by 2 PwC . f / H. f / jQC . f /j2 (8.527)
b) The second term is also a quadratic functional, with the difference with respect to the previous term that a transformation given by a periodic repetition is performed on H. f /; it can be proven that the gradient with respect to H. f / is C1 X 2 Pa . f / H f T2 `D 1 ` T (8.528)
c) The third term is a linear functional; from (8.491) the gradient is 2 Pa . f / T d) The fourth term is a constant; therefore the gradient vanishes. Considering that, if QC . f / is known, we obtain
G Rc . f /
(8.529)
min J D
H. f /DQC . f /G Rc . f /e j2 f t0
min
(8.530)
the optimum value of H. f / is obtained by setting rH J D 0 Extending the variational analysis to any h, we obtain rH J D 2 C1 X PwC . f / 2 H. f / C 2 Pa . f / H f jQC . f /j2 T `D 1 ` T 2 Pa . f / D 0 T (8.532) (8.531)
Now we introduce the signal G. f / that coincides with the rst term of (8.532), G. f / D PwC . f / H. f / jQC . f /j2 (8.533)
As both the second and the third term on the left-hand side of (8.532) are periodic functions of period 1=T , then also G. f / is a periodic function of period 1=T . Substituting therefore (8.533) in (8.532), it must be
C1 X 1 Pa . f / G T2 `D 1
G. f / C
QC f ` T PwC f
` 2 T ` T
1 Pa . f / D 0 T
(8.534)
739
Considering that the function G. f / is periodic, it can be brought out of the equation; after some steps we get the optimum solution for G. f /: G. f / D Pa . f / QC f C1 1 X T C Pa . f / T `D 1 PwC f ` 2 T ` T (8.535)
Finally, using (8.517), the expression of the optimum receive lter is given by G Rc . f / D
QC . f / e PwC . f / j2 f t0
Pa . f / QC f C1 X 1 T C Pa . f / T `D 1 Pw f C
` 2 T ` T
(8.537)
8.A.3
Referring to Figure 8.55, we remove the assumption that qC .t/ D h T x gC .t/ is known and reformulate the problem in the following terms. Given a transmission system with channel frequency response GC . f /, we wish to determine the frequency responses of the transmit lter HT x . f / (or equivalently QC . f / D HT x . f /GC . f /) and of the receive lter G Rc . f / that minimize the functional J given by (8.512). In this formulation we get a solution of interest only by setting a constraint on the statistical power of the transmitted signal, 2 Z C1 1 HT x . f / Pa . f / d f D V (8.538) MT x D T 1 However, even for the particular case of i.i.d. input symbols, for which Z 2 C1 MT x D a jHT x . f /j2 d f T 1
(8.539)
740
the problem cannot be solved in closed form. Here we give the procedure to determine the solution [58, 59]. 1. For each frequency f 2 [
1 2T 1 ; 2T ], determine the integer `max such that
` 2 T ` T
(8.540)
(8.541)
2. Choose a positive real number , then determine the subset of B dened as 1 jGC . f /j2 B D f 2 B : (8.542) > PwC . f / 3. Compute the magnitude of HT x . f /: 8 p < T PwC . f / 2 jHT x . f /j2 D : a jGC . f /j 0
f 2 B f 2 B =
(8.543)
jG Rc . f /j is determined using (8.537); the phase of HT x . f / and of G Rc . f / must be such that Q R . f / D HT x . f /GC . f /G Rc . f / is a positive real-valued signal. 4. Substituting (8.543) in (8.539), compute MT x . Verify whether MT x D V , otherwise steps 24 must be repeated for a different value of until MT x D V . As observed in [58], the physical interpretation of (8.543) is that HT x . f / has non-zero components in relation to frequencies at which the channel does not attenuate too much with respect to the noise level. However, this ratio is dened by 1=, where is a Lagrange multiplier used in the optimization of problem (8.512) under the constraint (8.539).
741
Appendix 8.B
8.B.1
We discuss a DFE design method that requires knowledge of the correlation sequences of the equalizer input signal and of the detected message, as well as their cross-correlation sequence [60]. With reference to Figure 8.13, we introduce the vectors whose elements are given by the coefcients of the feedforward and feedback lters, c D [c0 ; c1 ; : : : ; c M1
1] T
(8.544)
O D ; : : : ; a k D M2 ]
D ; : : : ; a k D M2 ]
(8.545)
O D 1 ; : : : ; a k D M2 ]
D 1 ; : : : ; a k D M2 ]
We note the presence of a delay D due to the feedforward lter. From (8.91), the sample at the decision point, yk , can be expressed using the following vector notation: Ok yk D cT xk C bT a0 Using (8.546), also the estimation error can be expressed in vector form as ek D yk ak
D
(8.546)
Ok D cT xk C bT a0
(8.547) ak
D
Ok Assuming that the detection of past symbols is correct, we have a0 D a0 . Then, introducing k Q (8.547) can be written as the vector b, ek D cT xk C [ 1; bT ]ak Q D c T xk C b T ak Q The lter coefcients c and b are determined by minimizing the cost function J D E[jyk ak
2 Dj ]
(8.548)
(8.549)
742
(8.550)
(8.551)
On the other hand, minimization of the MSE implies orthogonality of the error ek with respect to the input xk . Consequently, using the expression (8.548) for ek it must be
T Q E[ek xk ] D c H Rx C b H Rax D 0
(8.552)
or Q b H Rax D c H Rx (8.553)
Q b H Rax Rx 1
(8.554)
This relation, inserted in (8.551), yields JD Q Q Q b H Rax c C b H Ra b Q Q Q b H Rax Rx 1 Rxa b C b H Rax c (8.555)
Q Rax Rx 1 Rxa /b
Rajx D Ra
Rax Rx 1 Rxa
(8.556)
is the correlation matrix of the estimation error vector (see Appendix 2.A). Q The problem thus reduces to nding the vector b that minimizes the quadratic form Q Q J D b H Rajx b (8.557)
Q Q Dening e1 D [1; 01M2 ]T and remembering that b1 , the rst component of b, must be equal to 1, the solution is obtained by the minimization of the quadratic function Q Q J D b H Rajx b (8.558)
743
-1
b2
Figure 8.56. Constrained cost function in the case M2 D 2.
Q Figure 8.56 illustrates the problem in the case M2 D 2. The vector b that minimizes J subject to the constraint v D 0 is obtained by the method of Lagrange multipliers, ( rb J C rb v D 0 Q Q (8.560) vD0 Q As rb J D 2Rajx b and rb v D e1 , (8.560) yields Q Q ( Q 2Rajx b D e1 vD0 whose solution is given by Q bopt D 1 .Rajx /
1
(8.561)
e1
(8.562)
Q where is the factor that forces the rst component of bopt equal to 1. We note that, Q opt coincides with the rst column of the matrix .Rajx / 1 . apart from the coefcient 1=, b Q With this choice of bopt , (8.554) yields the optimum FF lter coefcients copt D Q Rx 1 Rxa bopt (8.563)
Although this procedure is very general, valid for a general statistic of the information message, it is rather computationally expensive because it requires two matrix inversions, for Rajx and Rx .
744
Observation 8.15 O For an LE, the formulation is as in (8.563), however, without the lter b, and with ak D Q D [ 1]. [ak D ] and b O
8.B.2
This method is obtained by applying a matrix formulation similar to that of Section 8.5. We introduce a denition that will simplify the notation. Denition 8.4 Let q D [q1 ; : : : ; q N ]T be a vector with N elements, and qi j D [qi ; qi C1 ; : : : ; q j ]T (8.564)
denote the vector containing a subsequence of consecutive elements of q. Let A D [An;m ], n D 1; : : : ; N R , m D 1; : : : ; NC , be a N R NC matrix. If A;m denotes the m-th column, then A;i j D [A;i ; A;i C1 ; : : : ; A; j ] (8.565)
denotes the matrix containing a subsequence of consecutive columns of the matrix A. Figure 8.57 illustrates the vector representation of the scheme of Figure 8.5, extended to include a DFE. Let fh i g, i D N1 ; : : : ; N2 , be the overall impulse response at the equalizer input. Introducing the vectors xk D [x k ; x k
1 ; : : : ; x k M1 C1 ] T N2 M1 C1 ] T T
ak D [akCN1 ; : : : ; ak ; : : : ; ak Q wk D [wk ; wk Q Q
(8.566)
Q 1 ; : : : ; wk M1 C1 ]
and using (8.544), it is possible to express the input vector xk of the FF lter as a linear combination of the transmitted symbols ak plus noise Q xk D Hak C wk (8.567)
745
where 2 6 6 6 6 HD6 6 6 4 h
N1
0 0 : : : 0
h N1 C1 h N1 C2 : : : h N2 0 0 0 h N1 h N1 C1 : : : h N2 1 h N2 0 h N1 : : : h N2 2 h N2 1 h N2 :: : 0 0 ::: h
N1
0 0 0 : : : h N2
3 7 7 7 7 7 7 7 5
(8.568)
N1 C1
is an M1 .N1 C N2 C M1 / Toeplitz matrix . We note that in (8.566) the denition of ak refers to the system memory at the output of the lter c: It is useful to express the operation of the FB lter assuming as an input the vector ak , given by (8.566). Therefore we introduce the vector b0 D [01.N1 CDC1/ ; bT ; 01M3 ]T where M3 D M1 C N2 D 1 M2 . The sample yk at the decision point can thus be expressed as Q Q yk D cT xk C b0 T ak D cT Hak C b0 T ak C cT wk D .cT H C b0 T /ak C cT wk (8.570) (8.569)
Equation (8.570) suggests that the cancellation of ISI from the sample yk , due to the symbols fak D i g, i D 1; : : : ; M2 , is obtained by setting
N b0 D [01.N1 CDC1/ ; .cT H/ N1 CDCM2 C1 ; 01M3 ]T 1 CDC2
(8.571)
(8.572)
2 6 6 D6 6 4
N1
0 : : : 0
h N1 C1 : : : h D h N1 : : : h D 1 : :: : : : 0 ::: 0
0 0 : : : 0 h N2
0 0 : : :
M3 C1
::: :::
0 0 : : :
3 7 7 7 7 5 (8.573)
: : : h N2
Q We are now back to the case of one lter c, having as input vector .H0 ak C wk /, whose coefcients are determined by minimizing the cost function J D E[jyk ak
2 Dj ]
(8.574)
746
From the Wiener theory, we know that to nd the solution we must compute the autocorrelation matrix R of the input signal and the cross-correlation vector p between the desired output signal and the input signal. From Q Q R D E[.H0 ak C wk / .H0 ak C wk /T ]
2 assuming the symbols of the sequence fak g are i.i.d. with variance a , we obtain 2 R D a H0 H0 T C Rw Q
(8.575)
(8.576)
where Rw is the autocorrelation matrix of the noise, assumed known. Q Moreover, we have p D E[ak
0 D .H ak
Q C wk / ] D H0 E[ak
D ak ]
(8.577)
given the statistical independence between transmitted symbols and noise. As the symbols are i.i.d., the vector E[ak D a ] has all zero elements except that in position N1 C D C 1, k 2 with value a , and p corresponds to the .N1 C DC1/-th column of the matrix H0 multiplied 2 by the scalar a ,
2 2 p D a H;N1 CDC1 D a [h D ; h D
0
(8.578)
According to the Wiener theory, the optimum FF lter coefcients are given by copt D R
1
p
1 2 0 a H;N1 CDC1
2 D .a H0 H0 T C Rw / Q
(8.579)
p H copt
(8.580)
The feedback lter is obtained by substituting (8.579) in (8.571). Observation 8.16 For an LE, the formulation is as in (8.579) and (8.580), however, without the lter b, and with H0 equal to H.
8.B.3
Method based on the channel impulse response and any symbol statistic
T Ra D E[a ak ] k T Rx D E[x xk ] D H Ra HT C Rw Q k T Rax D E[a xk ] D Ra HT k T Rxa D E[x ak ] D H Ra k
For the signal (8.567), the correlation matrices (8.550) are given by
(8.581)
747
Using (8.581) and (8.569) in (8.555), from (3.176) we obtain Q J D b H [Ra Q Rax Rx 1 Rxa ]b
1Q
Q D b H [Ra 1 C HT Rw 1 H ] Q Setting
(8.582) b
3 7 5
This quadratic form is minimized by choosing as FB lter coefcients (see (8.562)) [ 1; bopt;1 ; : : : ; bopt;M2 ]T D R D1 e1
T e1 R D1 e1
(8.585)
With this choice for the FB lter, the minimum value of the MSE is given by Jmin D 1
T e1 R D1 e1
(8.586)
and the optimum FF lter can be obtained by (8.563). Observation 8.17 We note that if the number of coefcients M2 of the FB lter is chosen equal to the length of the overall channel impulse response at the decision point, that is if M2 D N1 C N2 C 1, the performance is improved; moreover, it is possible to reduce the computational complexity by exploiting the special structure of the matrices [60].
8.B.4
FS-DFE
Although the FB lter necessarily operates with input samples having sampling period equal to T , the FF lter can be fractionally spaced. According to the Wiener theory, the optimum coefcients of c and b are obtained by one of the two methods presented in the previous sections; however, care must be taken in accounting for the fact that the FF lter operates with input samples having sampling period equal to T =F0 , where F0 is the oversampling factor.
748
iD
N1 ; : : : ; N 2
(8.587) (8.588)
Q Q 1/ ; : : : ; wk F0 C1 ; wk F0 ]
1/ ; : : : ; x k F0 ] T
N2 X i D N1
hi ak
Q C wk
(8.589)
(8.590)
(8.591)
The input signal to the lter c is still denoted by the vector xk , whose components are the vectors fxi g, i D k; : : : ; k M1 1, each with dimension F0 . Then we can write 3 2 xk 7 6 : : Q xk D 4 (8.592) 5 D Hak C wk : xk
M1 1
3 7 7 7 7 7 7 7 5
(8.593)
Note that the above relation is formally identical to (8.567). Therefore, thanks to this property, we can obtain the optimum coefcients of the FF and FB lters by the methods introduced in the previous sections.
749
Appendix 8.C
The equalization algorithm discussed in this appendix is related to the eye diagram at the decision point [4]. For an equalizer with N coefcients, let n be the overall impulse response at the decision point:
n
N 1 X jD0
c j hn
n
nD
N1 ; : : : ; N 2 C N
(8.594)
n D
nCD
(8.595)
such that 0 is the desired sample at the decision point. Typically, DD N 2 1 (8.596)
To simplify the notation, we assume that 0 coincides with the peak 0 D max jn j
n
(8.597)
the cost function J to be minimized considers the peak value of ISI, that is JD We note that J D 0 for ( n D 0 6D 0 0 nD0 n 6D 0 (8.601) 1 0
L2 X nD L 1 ; n6D0
jn j
(8.600)
and the system is equalized, hence c is a zero forcing lter. On the other hand, if J > 1, then the eye is completely shut. In any event, the effect of the noise is not considered. As illustrated in Figure 8.58, the equalizer coefcients are determined iteratively in relation to a repetitive input sequence, with repetition index i, composed of a symbol equal to one followed by L 1 zeros: fak ; k D 0; : : : ; L 1g D f1; 0; : : : ; 0g.
750
ak
i=0
i=1
i=2
0 1
(i) {h n }
(L1)
(i) {n }
j D 0; : : : ; N
N 1 X
j 6D D
(8.602) (8.603)
c.i C1/ h D j
jD0; j6D D
The last equation is obtained from the constraint 0 D 1. It can be proved that if the eye is open at the equalizer input, that is 1 h0
N2 X nD N1 ; n6D0
jh n j < 1
(8.604)
then (8.602) converges and (8.601) is veried. It is useful to observe that in the absence of noise it would be sufcient to send only one sequence fak g, k D 0; : : : ; L 1: in practice, the various estimates of fh n g are used for averaging with respect to noise. We note the simplicity of implementing this algorithm: indeed, it was one of the rst algorithms to equalize transmission lines. Now its use is justied only in systems with very large modulation rates, for example, optical ber systems. In any case, for the algorithm to converge, the eye needs to be already opened, a constraint that did not exist in the MSE criterion. Another distinct characteristic of this method derives from the fact that it uses test sequences fak g, different from those typically used for transmission.
751
Appendix 8.D
We consider a discrete-time system with input sequence fi k g and output sequence fok g, evaluated at instants kT: We say that the output sequence is generated by an FSM if there are a sequence of states fsk g and two functions f o and f s , such that ok D f o .i k ; sk sk D f s .i k ; sk
1/ 1/
(8.605) (8.606)
as illustrated in Figure 8.59. The rst equation describes the fact that the output sample depends on the current input and the state of the system. The second equation represents the memory part of the FSM and describes the state evolution. We note that if f s is a one-to-one function of i k , that is if every transition between two states is determined by a unique input value, then (8.605) can be written as ok D f .sk ; sk
1/
(8.607)
Chapter 9
For channels that exhibit high signal attenuation at frequencies within the passband, a valid alternative to CAP/QAM is represented by a modulation technique based on lter banks, known as orthogonal frequency division multiplexing (OFDM), or multicarrier modulation. As the term implies, multicarrier modulation is obtained in principle by modulating several carriers in parallel using blocks of symbols, therefore using a symbol period that is typically much longer than the symbol period of a single-carrier system transmitting at the same bit rate. The resulting narrowband signals around the frequencies of the carriers are then added and transmitted over the channel. The narrowband signals are usually referred to as subchannel signals. An advantage of OFDM with respect to single-carrier systems is represented by the lower complexity required for equalization, that under certain conditions can be performed by a lter with a single coefcient per subchannel. A long symbol period also yields a greater immunity of an OFDM system to impulse noise; however the symbol duration, and hence the number of subchannels, are limited for transmission over time-variant channels. As we will see in this chapter, another important aspect is represented by the efcient implementation of modulator and demodulator, obtained by sophisticated signal processing algorithms.1
9.1
OFDM systems
Orthogonal frequency division multiplexing is an efcient modulation technique by which blocks of M symbols are transmitted in parallel over M subchannels, using M modulation lters [1] with frequency responses Hi . f /, i D 0; : : : ; M 1. We consider the baseband equivalent OFDM system illustrated in Figure 9.1. The M input symbols at the k-th modulation interval are represented by the vector ak D [ak [0]; : : : ; ak [M 1]]T (9.1)
In this text, we refer to OFDM as a general technique for multicarrier transmission. In particular, this applies to both wired and wireless systems. When multicarrier transmission is achieved without ltering of the individual subchannel signals, other authors use the term discrete multitone (DMT) modulation for wired transmission systems, whereas they reserve the term OFDM for wireless systems.
754
where ak [i] 2 A[i], i D 0; : : : ; M 1. The alphabets A[i], i D 0; : : : ; M 1, correspond to two dimensional constellations, which are not necessarily identical. The symbol rate of each subchannel is equal to 1=T and corresponds to the rate of the input symbol vectors fak g: 1=T is also called modulation rate. The sampling period of each input sequence is changed from T to T =M by upsampling, that is by inserting M 1 zeros between consecutive symbols (see Appendix 1.A); each sequence thus obtained is then ltered by a bandlimited lter properly allocated in frequency. The lter on the i-th branch is characterized by the impulse response fh n [i]g, with transfer function Hi .z/ D
C1 X nD 1
h n [i] z
(9.2)
and frequency response Hi . f / D Hi .z/j zDe j2 f T =M , i D 0; : : : ; M 1. The transmit lters, also called interpolator lters, work in parallel at the rate M=T , called transmission rate. The transmitter output signal sn is given by the sum of the lter output signals, that is sn D
M 1 X X C1
i D0 kD 1
ak [i] h n
k M [i]
(9.3)
The received signal rn , suitably delayed of D0 samples, is ltered by M lters in N parallel at the transmission rate. The receive lters have impulse responses fgn [i]g, i D P 0; : : : ; M 1, and transfer functions given by G i .z/ D n gn [i]z n , i D 0; : : : ; M 1. By downsampling the M output signals of the receive lters at the modulation rate 1=T , we obtain the vector sequence fyk D [yk [0]; : : : ; yk [M 1]]T g, which is used to detect the vector sequence fak g; receive lters are also known as decimator lters. When the channel is ideal and noiseless, we get rn D sn . The delay D0 is chosen to obtain proper N synchronization of the receive lter bank. We consider as transmit lters FIR causal lters having length M, and support f0; 1; : : : ; M 1g, that is h n [i] may be non-zero for n D 0; : : : ; M 1. We assume matched receive lters; as we are dealing with causal lters, we consider the expression of
755
1, given by
n [i]
8n
(9.4)
We observe that the support of fgn [i]g is f1; 2; : : : ; Mg, and, for an ideal channel, D0 D 0.
9.2
Orthogonality conditions
Assuming the channel is ideal and noiseless, if the transmit and receive lter banks are designed such that certain orthogonality conditions are satised, the subchannel output signals are delayed versions of the transmitted symbol sequences at the corresponding subchannel inputs. The orthogonality conditions, also called perfect reconstruction conditions, can be interpreted as a more general form of the Nyquist criterion.
Time domain
With reference to the general scheme of Figure 9.1, at the output of the j-th receive subchannel, before downsampling, the impulse response relative to the i-th input is given by
M 1 X
pD0
h p [i] gn
p [ j] D
M 1 X
pD0
h p [i] h MC p
n [ j]
8n
(9.5)
We note that, for j D i, the peak of the sequence given by (9.5) is obtained for n D M. Observing (9.5), transmission in the absence of intersymbol interference over a subchannel, as well as absence of interchannel interference (ICI) between subchannels, is achieved if orthogonality conditions are satised, that in the time domain are expressed as
M 1 X
pD0
h p [i] h M. pC
k/ [ j]
D i
j k
i; j D 0; : : : ; M
(9.6)
Hence, in the ideal channel case considered here, the vector sequence at the output of the decimator lter bank is a replica of the transmitted vector sequence with a delay of modulation intervals, that is fyk g D fak g. Sometimes the elements of a set of orthogonal impulse responses that satisfy (9.6) are called wavelets.
Frequency domain
In the frequency domain, the conditions (9.6) are expressed as
M 1 X
`D0
Hi
M ` Hj f T
M ` T
D i
i; j D 0; : : : ; M
(9.7)
z-transform domain
In the ideal channel case, the relation between the inputs ak [i], i D 0; : : : ; M 1, and the j-th output yk [ j] is represented by the block diagram of Figure 9.2. We note that every
756
a k [0]
T
a k [1]
T M T M
H0 (z) H1 (z)
Gj (z)
T M T M
Gj (z)
. . .
M
. . .
y k [j]
a k [M-1]
T M
H M
-1
(z)
Gj (z)
T M
Figure 9.2. Block diagram that illustrates the relation between the M inputs and the j-th output.
branch in the diagram represents a time-invariant linear system. Setting a.z/ D " X
k
ak [0]z
;:::;
X
k
#T ak [M 1]z
k
(9.8)
and " X
k k
y.z/ D
yk [0]z
;:::;
X
k
#T yk [M 1]z
k
(9.9)
in general the inputoutput relation is given by (see Figure 9.6) y.z/ D S.z/ a.z/ (9.10)
where the element [S.z/] p;q of the matrix S.z/ is the transfer function between the q-th input and the p-th output. We note that the orthogonality conditions, expressed in the time domain by (9.6), are satised if S.z/ D z where I is the M M identity matrix. I (9.11)
9.3
For large values of M, a direct implementation of an OFDM system as shown in Figure 9.1 would require an exceedingly large computational complexity, as all ltering operations are performed at a high rate equal to M=T . A reduction in the number of operations per unit of time is obtained by resorting to the polyphase representation of the various lters [2].
757
E i.`/ .z M /
i D 0; : : : ; M
(9.12)
where E i.`/ .z/ is the transfer function of `-th polyphase component of Hi .z/. We dene the vector of the transfer functions of transmit lters as h.z/ D [H0 .z/; : : : ; HM Let e.z/ be the vector of delay elements given by e.z/ D [z .M
1/ 1 .z/] T
(9.13)
; z .M
2/
; : : : ; 1]T
and E.z/ the matrix of the transfer functions of the polyphase components lters, given by 2 .0/ .0/ .0/ E 0 .z/ E 1 .z/ E M 1 .z/ 6 .1/ .1/ .1/ 6 E 1 .z/ E M 1 .z/ E 0 .z/ 6 E.z/ D 6 6 : : : : : : 6 : : : 4 . . .M E 0M 1/ .z/ E 1M 1/ .z/ E M 11/ .z/ From the identity z .M h.z/ can be expressed as hT .z/ D z .M
1/ H 1/ H
(9.15)
1 z
D [1; z
; : : : ; z .M
1/
(9.16)
1 z
E.z M /
(9.17)
Observing (9.17), the M interpolated input sequences are ltered by lters whose transfer functions are represented in terms of the polyphase components, expressed as columns of the matrix E.z M /. Because the transmitted signal fsn g is given by the sum of the lter outputs, the operations of the transmit lter bank are illustrated by the scheme of Figure 9.3, where the signal fsn g is obtained from a delay line that collects the outputs of the M vectors of lters with transfer functions given by the rows of E.z M /, and input given by the vector of interpolated input symbols. At the receiver, using a representation equivalent to (1.637), that is obtained by a permutation of the polyphase components, we get G i .z/ D
M 1 X
`D0
z .M
1 `/
Ri.`/ .z M /
i D 0; : : : ; M
(9.18)
758
where Ri.`/ .z/ is the transfer function of the .M 1 `/-th polyphase component of G i .z/. Then the vector of the transfer functions of the receive lters g.z/ D [G 0 .z/; : : : ; G M 1 .z/] T can be expressed as g.z/ D R.z M / e.z/ (9.19)
where R.z/ is the matrix of the transfer functions of the polyphase components of the receive lters, and is given by 2 3 .0/ .1/ . R 0 .z/ R 0M 1/ .z/ R0 .z/ 6 7 .0/ .1/ . 6 R1 .z/ R 1 .z/ R 1M 1/ .z/ 7 6 7 R.z/ D 6 (9.20) 7 : : : 6 7 : : : : : : 4 5
.0/ RM 1 .z/ .1/ RM 1 .z/
.M R M 11/ .z/
Observing (9.19), the M signals at the output of the receive lter bank before downsampling are obtained by ltering in parallel the received signal by lters whose transfer functions are represented in terms of the polyphase components, expressed as rows of the matrix R.z M /. Therefore the M output signals are equivalently obtained by ltering the N N vector of received signal samples [Nn .M 1/ ; rn MC2 , : : : ; rn ]T by the M vectors of lters r with transfer functions given by the rows of the matrix R.z M /. In particular, recalling that the receive lters have impulse responses given by (9.4), we obtain 1 0i M 1 (9.21) G i .z/ D z M Hi z Substituting (9.21) in (9.19), and using (9.17), the expression of the vector g.z/ of the transfer functions of the receive lters becomes 1 1 g.z/ D z M h H D z MCM 1 E H e.z/ (9.22) z z M
759
a k [0] T
yk [0] T
a k [1] T r = sn n T M r n
0
yk [1] T
E(z)
. . .
P/S
-D
S/P
(1)
E 1 z*
( )
. . .
M a [ -1]
k
yk M -1] [ T
a)
b)
Figure 9.5. Block diagrams of (a) parallel to serial converter, (b) serial to parallel converter.
Apart from a delay term z 1 , the operations of the receive lter bank are illustrated by the scheme of Figure 9.3. From Figure 9.3 using (9.16), (9.22), and applying the noble identities given in Figure 1.70, we obtain the system represented in Figure 9.4, where parallel to serial (P/S) and serial to parallel (S/P) converters are illustrated in Figures 9.5a and 9.5b, respectively.
760
y k[0]
y k[1]
a k[ T
1]
y k[ T
1]
Figure 9.6. Equivalent OFDM system with inputoutput relation expressed in terms of the matrix S.z/.
From (9.11), in terms of the polyphase components of the transmit lters in (9.15), the orthogonality conditions in the z-transform domain are therefore expressed as EH 1 z E.z/ D I (9.24)
Equivalently, using the polyphase representation of the receive lters in (9.20), we nd the conditions 1 H R.z/ R DI (9.25) z
In other words, the spacing in frequency between the subcarriers is 1 f D 1=T . To derive an efcient implementation of uniform lter banks we consider the scheme represented in Figure 9.7a, where H .z/ and G.z/ are the transfer functions of the transmit and receive prototype lters, respectively. We consider as transmit prototype lter a causal FIR lter with impulse response fh n g, having length M and support f0; 1; : : : ; M 1g; the receive prototype lter is a causal FIR lter matched to the transmit prototype lter, with impulse response given by gn D h M
n
8n
(9.27)
761
Figure 9.7. Block diagram of an OFDM system with uniform lter banks: (a) general scheme, (b) equivalent scheme for fi D i=T, i D 0; : : : ; M 1.
With reference to Figure 9.7a, the i-th subchannel signal at the channel input is given by sn [i] D e j2 M As e j2 M D e j2 we obtain sn [i] D where h n [i] D h n e j2 M Recalling the denition WM D e
2 jM in 1 X kD 1 in i .n k M/ in 1 X kD 1
ak [i] h n
kM
(9.28)
(9.29)
1 X
ak [i] h n
j2 kM e
i .n k M/
ak [i] h n
k M [i]
(9.30)
kD 1
(9.31)
i Hi .z/ D H .zW M /
Observing (9.30) and (9.32), we obtain the equivalent scheme of Figure 9.7b.
762
The scheme of Figure 9.7b may be considered as a particular case of the general OFDM scheme represented in Figure 9.1; in particular, the transfer functions of the lters can be expressed using the polyphase representations, which are, however, simplied with respect to the general case expressed by (9.15) and (9.20), as we show in the following. Observing (9.28) we express the overall signal sn as sn D
M 1 X
i D0
e j2 M
in
C1 X kD 1
hn
k M ak [i]
e j2 M .m MC`/
C1 X kD 1
h .m
k/MC` ak [i]
(9.34)
.`/ Observing that e j2i m D 1, setting sm D sm MC` and h .`/ D h m MC` , and interchanging m the order of equations, we express the `-th polyphase component of fsn g as .`/ sm D C1 X kD 1
h .`/ k m
M 1 X
i D0
i WM` ak [i]
(9.35)
The sequences fh .`/ g, ` D 0; : : : ; M 1, denote the polyphase components of the prototype m lter impulse response, with transfer functions given by H .`/ .z/ D
1 X mD0
h .`/ z m
` D 0; : : : ; M
(9.36)
Recalling the denition (1.94) of the DFT operator as M M matrix, the IDFT of the vector ak is expressed as
1 FM ak D Ak D [Ak [0]; : : : ; Ak [M
1]]T
(9.37)
We nd that the inner summation in (9.35) yields X 1 M 1 i WM` ak [i] D Ak [`] M i D0 and
.`/ sm D 1 X kD 1
` D 0; 1; : : : ; M
(9.38)
M h .`/ k Ak [`] D m
1 X pD 1
M h .`/ Am p
p [`]
(9.39)
Including the factor M in the denition of the prototype lter impulse response, or in M gain factors that establish the statistical power levels to be assigned to each subchannel signal, an efcient implementation of a uniform transmit lter bank is given by an IDFT, a polyphase network with M branches, and a P/S converter, as illustrated in Figure 9.8.
763
a k [0] T
Ak [0]
H (z)
(0)
sk
(0)
T
(1)
a k [1] T
Ak [1]
H (z)
(1)
sk
IDFT
P/S
sn T M
rn
GC(z)
w n
a k [M -1] T
Ak [ -1] M
(M -1)
(z)
sk
(M-1)
rk T
(0)
y [0]
k
^ [0] ak
T
rk rn
(1)
y k [1] T . . .
^ [1] ak
T
z -D 0
rn
S/P
. . .
rk
(M-1)
(M-1)* 1 z (1) H z*
yk [ M -1] T
^ [M -1] ak
T
Observing (9.35), we nd that the matrix of the transfer functions of the polyphase components of the transmit lters is given by E.z/ D diagfH .0/ .z/; : : : ; H .M
1/ 1 .z/gF M
(9.40)
Therefore the vector of the transfer functions of the transmit lters is expressed as 1 1 T .M 1/ H h .z/ D z diagfH .0/ .z M /; : : : ; H .M 1/ .z M /gFM e (9.41) z We note that we would arrive at the same result by applying the notion of prototype lter with the condition (9.26) to (9.15). The vector of the transfer functions of a receive lter bank, which employs a prototype lter with impulse response given by (9.27), is immediately obtained by applying (9.22), with the matrix of the transfer functions of the polyphase components given by (9.40). Therefore we get 1 1 ; : : : ; H .M 1/ e.z/ (9.42) g.z/ D z MCM 1 FM diag H .0/ zM zM
764
Hence an efcient implementation of a uniform receive lter bank, also illustrated in Figure 9.8, is given by a S/P converter, a polyphase network with M branches, and a DFT; we note that the lter of the i-th branch at the receiver is matched to the lter of the corresponding branch at the transmitter. With reference to Figure 9.7a, it is interesting to derive (9.42) by observing the relation N between the received sequence rn D rn D0 and the output of the i-th subchannel yk [i], given by yk [i] D
C1 X nD 1
gk M
i T j2 T n M rn
(9.43)
With the change of variables n D mM C `, for m D 1; : : : ; C1, and ` D 0; : : : ; M 1, and recalling the expression of the prototype lter impulse response given by (9.27), we obtain yk [i] D
i
M 1 X X C1
`D0 mD 1
h .
kCm/MC`
i j2 M .m MC`/
rm MC`
(9.44)
.`/ Observing that e j2 M m M D 1, setting rm D rm MC` , and h .`/ D h MC` , and interm m changing the order of equations, we get
yk [i] D
2 j M i`
M 1 X
`D0
2 j M i`
1 X mD 1
h .`/ Cm
.`/ k rm
(9.45)
yk [i] D
M 1 X
`D0
i` WM
1 X mD 1
h .`/ Cm
.`/ k rm
(9.46)
Provided the orthogonality conditions are satised, from the output samples yk [i], i D 0; : : : ; M 1, threshold detectors may be employed to yield the detected symbols ak [i], O i D 0; : : : ; M 1, with a certain delay D. As illustrated in Figure 9.8, all ltering operations are carried out at the low rate 1=T . Also note that in practice the FFT and the inverse FFT are used in place of the DFT and the IDFT, respectively, thus further reducing the computational complexity.
9.4
For the above discussion on the efcient realization of OFDM systems, we referred to Figure 9.8, where the number of subchannels, M, coincides with the interpolation factor of the modulation lters in the transmit lter bank. These systems are called lter bank systems with critical sampling or, in short, critically sampled lter banks. We now examine the general case of M modulators where each use an interpolation lter by a factor K > M: this system is called non-critically sampled. In principle the schemes
765
Figure 9.9. Block diagram of (a) transmitter and (b) receiver in a transmission system employing non-critically sampled lter banks, with K > M and fi D .iK/=.MT/ D i=.MTc /.
of transmit and receive non-critically sampled lter banks are illustrated in Figure 9.9. As in critically sampled systems, also in non-critically sampled systems it is advantageous to choose each transmit lter as the frequency-shifted version of a prototype lter with impulse response fh n g, dened over a discrete-time domain with sampling period Tc D T =K. At the receiver, each lter is the frequency-shifted version of a prototype lter with impulse response fgn g, also dened over a discrete-time domain with sampling period Tc D T =K. As depicted in Figure 9.10, each subchannel lter has a bandwidth equal to K=.MT /, larger than 1=T . Maintaining a spacing between subcarriers of 1 f D K=.MT /, it is easier to avoid spectral overlapping between subchannels and consequently to avoid ICI. It is also possible to choose fh n g, e.g., as the impulse response of a square root raised cosine lter, such that, at least for an ideal channel, the orthogonality conditions are satised and ISI is also avoided. We note that this advantage is obtained at the expense of a larger
766
bandwidth required for the transmission channel, that changes from M=T for critically sampled systems to K=T for non-critically sampled systems. Therefore the system requires an excess bandwidth given by .K M/=M. Also for non-critically sampled lter banks it is possible to obtain an efcient implementation using the discrete Fourier transform [3, 4]. The transmitted signal is expressed as a function of the input symbol sequences as sn D or, equivalently, sn D With the change of indices n D mM C ` (9.48) becomes sm MC` D
1 X kD 1 1 X kD 1
M 1 X
i D0
e j2 MT n K
iK
1 X kD 1
ak [i] h n
kK
(9.47)
hn
kK
M 1 X
i D0
in ak [i] WM
(9.48)
m2Z
` D 0; 1; : : : ; M
M 1 X
i D0
(9.49)
h m M k KC`
i WM` ak [i]
(9.50)
Using the denition of the IDFT (9.38), apart from a factor M that can be included in the impulse response of the lter, and introducing the following polyphase representation of the transmitted signal
.`/ sm D sm MC`
(9.51)
we obtain
.`/ sm D 1 X kD 1
h m M k KC` Ak [`]
(9.52)
By analogy with (1.561), (9.52) is obtained by interpolation of the sequence fAk [`]g by a factor K, followed by decimation by a factor M. From (1.569) and (1.570), we introduce the change of indices mM k (9.53) pD K
767
mM K
mM K
.mM/mod K K
(9.54)
h . pC1m /KC` Aj mM k
K
[`]
D Letting
h pKC`C.m M/mod K Aj mM k
K
[`]
p; m 2 Z
` D 0; 1; : : : ; M
(9.55)
h .`/ Aj mM k p;m
K
[`]
(9.56)
The efcient implementation of the transmit lter bank is illustrated in Figure 9.11. We note that the system is now periodically time-varying, i.e. the impulse response of the lter components cyclically changes. The M elements of an IDFT output vector are input to M delay lines. Also note that within a modulation interval of duration T , the samples stored in some of the delay lines are used to produce more than one sample of the transmitted signal. Therefore the P/S element used for the realization of critically sampled lter banks needs to be replaced by a commutator. At instant nT =K, the commutator is linked to the ` D n mod M -th ltering element. The transmit signal sn is then computed by
Figure 9.11. Efcient implementation of the transmitter of a system employing non-critically sampled lter banks; the lter components are periodically time-varying.
768
convolving the signal samples stored in the `-th delay line with the n mod K -th polyphase component of the T =K-spaced-coefcients prototype lter. In other terms, each element of the IDFT output frame is ltered by a periodically time-varying lter with period equal to [l :c:m:.M; K/]T =K, where l :c:m:.M; K/ denotes the least common multiple of M and K. Likewise, the non-critically sampled lter bank at the receiver can also be efciently implemented using the DFT. In particular, we consider the case of downsampling of the subchannel output signals by a factor K=2, which yields samples at each subchannel output at an (over)sampling rate equal to 2=T . With reference to Figure 9.9b, we observe that the output sequence of the i-th subchannel is given by yn 0 [i] D where gn D h M n . With the change of indices n D mM C ` m2Z ` D 0; 1; : : : ; M ! 1 (9.58)
1 X nD 1
gn 0 K
2
in j2 M
rn
(9.57)
yn 0 [i] D
M 1 X
`D0
1 X mD 1
gn 0 K
2
.`/ m M ` rm
i` WM
(9.59)
We note that in (9.59) the term within parenthesis may be viewed as an interpolation by a factor M followed by a decimation by a factor K=2. Letting 0 nK m (9.60) qD 2M and 1n 0 D n0K 2M n0K 2M D .n 0 K=2/mod M M (9.61)
gq MC.n 0 K /mod M
2
`r
.`/ j 0
n K=2 M
(9.62)
q; n 0 2 Z
` D 0; 1; : : : ; M
(9.63)
(9.64)
769
Figure 9.12. Efcient implementation of the receiver of a system employing non-critically sampled lter banks; the lter components are periodically time-varying (see (9.63)).
(9.65)
The efcient implementation of the receive lter bank is illustrated in Figure 9.12, where we assume for the received signal the same sampling rate of K=T as for the transmitted signal, and a downsampling factor K=2, so that the samples at each subchannel output are obtained at a sampling rate equal to 2=T . Note that the delay element z D0 at the receiver input has been omitted, as the optimum timing phase for each subchannel can be recovered by using per-subchannel fractionally spaced equalization, as discussed in Section 8.4 for single-carrier modulation. Also note that within a modulation interval of duration T , more than one sample is stored in some of the delay lines to produce the DFT input vectors. Therefore the S/P element used for the realization of critically sampled lter banks needs to be replaced by a commutator. After the M elements of a DFT input vector are produced, the commutator is circularly rotated K=2 steps clockwise from its current position, allowing a set of K=2 consecutive received samples to be input into the delay lines. The content of each delay line is then convolved with one of the M polyphase components of the T =K-spaced-coefcients receive prototype lter. A similar structure is obtained if in general a downsampling factor K0 K is considered.
9.5
We consider three simple examples of critically sampled lter bank modulation systems. For practical applications, equalization techniques and possibly non-critically sampled lter bank realizations are required, as will be discussed in the following sections.
770
a k [1] T
Ak [1]
y k [1] T
. . .
IDFT
P/S
sn T M
GC(z)
w n
rn
-D
rn
0
S/P
DFT
. . .
a k [M -1] T
Ak [ -1] M
rn( M -1)
yk [ M -1] T
Figure 9.13. Block diagram of an OFDM system with impulse response of the prototype lter given by a rectangular window of length M.
771
Figure 9.14. Amplitude of the frequency responses of adjacent subchannel lters in a DMT system for f 2 0; 0:06 M and M D 64. [From [4], c 2002 IEEE] T
j2 f T
if j f j
1 2T otherwise
(9.69)
where the parameter 0 1 controls the spectral roll-off of the lter. The frequency response exhibits spectral nulls at the band edges and, when used as the prototype lter characteristic, leads to transmission free of ICI but with ISI within a subchannel. For ! 1, the frequency characteristic of each subchannel exhibits steep roll-off towards the band edge frequencies. On the other hand, for D 0 the partial-response class I characteristic is obtained. In general, it is required that at the output of each subchannel the ICI is negligible with respect to the noise. The amplitude of the frequency responses ofsubchannel lters obtained with a minimumphase prototype FIR lter for f 2 0; 0:06 M , and design parameters M D 64, D 10, T and D 0:1 are illustrated in Figure 9.15.
772
Figure 9.15. Amplitude of the frequency responses of adjacent subchannel lters in an FMT system for f 2 0; 0:06 M , and design parameters M D 64, D 10, and D 0:1. [From [4], T c 2002 IEEE]
Figure 9.16. Amplitude of frequency responses of adjacent subchannel lters in a DWMT system for f 2 0; 0:06 M and M D 64. [From [4], c 2002 IEEE] T
773
of individual subchannel signals as compared to DMT. The orthogonality conditions are satised; each subchannel, however, requires a bandwidth larger than 1=T . In DWMT modulation, all signal processing operations involve real signals. Therefore, for the same number of dimensions per modulation interval of the transmitted signal, the minimum bandwidth of a subchannel for DWMT is half the minimum bandwidth of a subchannel for DMT or FMT modulation. The implementation of lter banks for DWMT is examined in Section 9.9.
9.6
774
root raised cosine lter, with Nyquist frequency equal to M=.2T /, is selected to implement the interpolator lter and the anti-aliasing lter. We recall now the conditions to obtain a real-valued transmitted signal fsn g. For OFDM systems with the efcient implementation illustrated in Figure 9.8, it is sufcient that the coefcients of the prototype lter and the samples Ak [i], 8k, i D 0; : : : ; M 1, are realvalued. Observing (9.38), the latter condition implies that the following Hermitian symmetry conditions must be satised M 2< ak ak [0] 2 (9.70) M [M ak [i] D ak i] i D 1; : : : ; 1 2 In this case, the symmetry conditions (9.70) also allow a further reduction of the implementation complexity of the IDFT and DFT. When fsn g is a complex signal, the scheme of Figure 9.18 is adopted, where the lters gT x and g Rc have the characteristics described above and f 0 is the carrier frequency. We note that this scheme is analogous to that of a QAM system, with the difference that the transmit and receive lowpass lters, with impulse responses gT x and g Rc , respectively, have different requirements. The baseband equivalent scheme shown in Figure 9.18b is obtained from the passband scheme by the method discussed in Chapter 7. As the receive lters approximate an ideal lowpass lter with bandwidth M=.2T /, the signal-to-noise ratio 0 at the channel output is assumed equal to that obtained at the output of the baseband
gTx gRc cos(2f0 t) /2 -sin(2f0 t) Im[.] DAC gTx gCh /2 -sin(2f0 t) gRc ADC T M
(a) General scheme.
Re[.] sn T M
DAC
w(t)
ADC T M rn T M
cos(2f0 t)
sn
T M
gC
C, n
rn
T M
wn
(b) Baseband equivalent scheme.
Figure 9.18. (a) Passband analog OFDM transmission scheme; (b) baseband equivalent scheme.
775
Figure 9.19. Block diagram of a DMT system with cyclic prex and frequency-domain equalizer.
776
the P/S conversion, where the Nc 1 samples of the cyclic extension are the rst to be sent, the Nc 1 C M samples are transmitted over the channel. At the receiver, blocks of samples of length Nc 1 C M are taken; the boundaries between blocks are set so that the last M samples depend only on the elements of only one cyclically extended block of samples. The rst Nc 1 samples of a block are discarded. We now recall the result (1.116). The vector rk of the last M samples of the block received at the k-th modulation interval is expressed as rk D
k gC
C wk
(9.72)
where gC D [gC;0 ; : : : ; gC;Nc 1 ; 0; : : : ; 0]T is the M-component vector of the channel impulse response extended with M Nc zeros, wk is a vector of additive white Gaussian noise samples, and k is an M M circulant matrix, given by 2 3 Ak [0] Ak [M 1] Ak [1] 6 Ak [1] Ak [0] Ak [2] 7 6 7 6 7 (9.73) k D6 : : : 7 : : : 4 5 : : : Ak [M 1] Ak [M 2] Ak [0] Equation (9.72) is obtained by observing that only the elements of the rst Nc columns of the matrix k contribute to the convolution that determines the vector rk , as the last M Nc elements of gC are equal to zero. The elements of the last M Nc columns of the matrix k are chosen so that the matrix is circulant, even though they might have been chosen arbitrarily. Moreover, we observe that the matrix k , being circulant, satises the relation 2 3 ak [0] 0 0 6 0 7 0 ak [1] 6 7 1 7 D diagfak g FM k FM D 6 (9.74) : : : 6 7 : : : 4 5 : : : 0 Dening the DFT of the vector gC as GC D [GC;0 ; GC;1 ; : : : ; GC;M
1] T
ak [M
1] (9.75)
D FM gC
and using (9.74), we nd that the demodulator output is given by xk D FM rk D diagfak g GC C Wk (9.76)
where Wk D FM wk is given by the DFT of the vector wk . Recalling the properties of wk , Wk is a vector of independent Gaussian r.v.s. Equalizing the channel using the zero-forcing criterion, the signal xk (9.76) is multiplied by the diagonal matrix K, whose elements on the diagonal are given by2 Ki D [K]i;i D 1 GC;i i D 0; 1; : : : ; M 1 (9.77)
To be precise, the operation indicated by (9.77), rather than equalizing the signal, that is received in the absence of ISI, normalizes the amplitude and adjusts the phase of the desired signal.
777
Therefore the input to the data detector is given by yk D Kxk D ak C KWk (9.78)
We assume that the sequence of input symbol vectors fak g is a sequence of i.i.d. random vectors. Equation (9.78) shows that the sequence fak g can be detected by assuming transmission over M independent and orthogonal subchannels in the presence of additive white Gaussian noise. A drawback of this simple equalization scheme is the reduction in the modulation rate by a factor .M C Nc 1/=M. Therefore it is essential that the length of the channel impulse response is much smaller than the number of subchannels, so that the reduction of the modulation rate due to the cyclic extension can be considered negligible. To reduce the length of the channel impulse response one approach is to equalize the channel before demodulation [8, 10, 11]. With reference to Figure 9.18b, a linear equalizer with input rn is used; it is usually chosen as the FF lter of a DFE that is determined by imposing a prexed length of the feedback lter, smaller than the length of the cyclic prex.
gn M
n 1 [i]
Nc 1 X n 2 D0
hn1
n 2 [i] gC;n 2
n2Z
(9.79)
In the given scheme, the M DFEs depend on the transmission channel. If the transmission channel is time variant, each DFE must be able to track the channel variations, or it must be recomputed periodically. Error propagation inherent to decision-feedback equalization can be avoided by resorting to precoding techniques, as discussed in Chapter 13. The application of precoding techniques in conjunction with trellis-coded modulation (TCM) for FMT transmission is addressed in [4]. Per-subchannel T -spaced equalization. We consider now an FMT system with critically sampled lter banks, and subchannel output signals obtained at the sampling rate of 1=T . The high level of spectral containment of the transmit lters suggests that, if the number
778
Figure 9.20. Per-subchannel equalization for an FMT system with non-critically sampled lter banks.
of subchannels is sufciently high, and the group delay in the passband of the transmission channel is approximately constant, the frequency response of every subchannel becomes approximately a constant. In this case, the effect of the transmission channel is that of multiplying every subchannel signal by a complex value. Therefore, as for DMT systems with cyclic prex, equalization of the transmission channel can be performed by choosing a suitable constant for every subchannel. We note, however, that, whereas for a DMT system with cyclic prex the model of the transmission channel as a multiplicative constant for each subchannel is exact if the length of the cyclic prex is larger than the length of the channel impulse response, for an FMT system such a model is valid only as an approximation. The degree of the approximation depends on the dispersion of the transmission channel and on the number M of subchannels. Assuming a constant frequency response for transmission over each subchannel, the equalization scheme is given in Figure 9.21a, where K i is dened in (9.77), and the DFE is designed to equalize only the cascade of the transmit and receive lters. Using (9.27) and (9.31) we nd that the convolution of transmit and receive lters is independent of the subchannel index: in fact, we obtain
1 X n 1 D0
h n 1 [i] gn M
n 1 [i]
1 X n 1 D0
h n 1 gn M
n1
D h eq;n
(9.80)
In this case, all DFEs are equal. Simplied per-subchannel T -spaced equalization. A further simplication is obtained by using the implementation of Figure 9.21b. The idea is that, in the presence of a transmission channel with at frequency response for each subchannel, a reconstruction of the signal is achieved by designing the `-th polyphase component of the receive prototype lter, g .`/ , to equalize the corresponding polyphase component h .`/ of the transmit prototype lter. In
779
Figure 9.21. (a) Equalization scheme for FMT in the case of approximately constant frequency response for transmission over each subchannel; (b) simplied scheme.
general, a DFE scheme can be used, where the `-th polyphase components of the receive lters, g .`/ and g .`/ , equalize the corresponding i-th polyphase component h .`/ of the FF FB overall subchannel impulse response.
9.7
Various algorithms may be applied to achieve synchronization of OFDM systems for transmission over dispersive channels, depending on the system and on the type of equalization adopted. For DMT systems two synchronization processes are identied: synchronization of the clock of the A/D converter at the receiver front-end, or clock synchronization, and synchronization of the vector rk at the output of the S/P element, or frame synchronization.
780
Clock synchronization guarantees alignment of the timing phase at the receiver with that at the transmitter; frame synchronization, on the other hand, extracts from the sequence of received samples the blocks of M C Nc 1 samples that form the received frames, and determines the boundaries of the sequence of vectors rk that are presented at the input of the DFT. In principle, for a channel input sequence given by s0 D 1, and sn D 0, n 6D 0, the channel impulse response of length Nc must appear in the rst Nc positions of the receive vector (see Figure 9.19). For the initial convergence of both synchronization processes, training sequences without cyclic prex are usually employed [12]. For FMT systems with non-critically sampled lter banks and fractionally-spaced equalization, the synchronization is limited to clock recovery (see Section 8.4).
9.8
For a passband OFDM transmission system, the signal fsn g is in general complex valued and is shifted to passband adopting for example the scheme illustrated in Figure 9.18.
Figure 9.22. Block diagram of an SSB modulator for a passband DWMT signal.
781
roll-off around DC we obtain a vestigial side band (VSB) modulator.3 However, because of the difcult recovery of the phase and frequency of the carrier, digital transmission systems using SSB and VSB modulators are characterized by lower performance as compared to systems that consider transmission of the double-sided signal spectrum. To overcome this difculty and preserve the spectral efciency of the transmission scheme, a pilot tone may be used to provide the required information for the carrier recovery. The transmission of pilot tones, however, does not represent in many cases a practical solution, as it reduces the power efciency of the system and introduces one or more spectral lines in the signal spectrum.
SSB and VSB modulations are used, for example, for the analog transmission of video signals and can also be considered for digital communication systems.
782
In practice, however, OFDM systems offer some considerable advantages with respect to CAP/QAM systems. OFDM systems achieve higher spectral efciency if the channel frequency response exhibits large attenuations at frequencies within the passband. In fact, the band used for transmission can be varied by increments equal to the modulation rate 1=T Hz, and optimized for each channel. Moreover, if the noise exhibits strong components in certain regions of the spectrum, the total band can be subdivided in two or more sub-bands. OFDM systems guarantee a higher robustness with respect to impulse noise. If the average arrival rate of the pulses is lower than the modulation rate, the margin against the impulse noise is of the order of 10 log10 .M/ dB. For typical values of M, OFDM systems achieve the same performance as QAM systems with a complexity that can be considerably lower. In multiple-access systems, the ner granularity of OFDM systems allows a greater exibility in the spectrum allocation. On the other hand, OFDM systems present also a few drawbacks with respect to QAM systems. In OFDM systems the transmitted signals exhibit a higher peak-to-average power ratio, that contributes to an increase in the susceptibility of these systems to nonlinear distortion. Because of the block processing of samples, a higher latency is introduced by OFDM systems in the transmission of information.
9.9
DWMT modulation
In DMT systems the reduction of the modulation rate, equal to the factor .MC Nc 1/=M, and the consequent need to reduce this penalty by shortening the impulse response of the transmission channel are due to the non-negligible spectral overlap of signals on adjacent subchannels. As a cyclic prex of length Nc 1 is used for equalization, the orthogonality between signals on different subchannels is veried only for channels with length of the impulse response smaller than or equal to Nc . Ideally, if the frequency response of the prototype lter is chosen such that signals of different subchannels do not overlap, the orthogonality is maintained independently of the impulse response of the transmission channel. Therefore it is interesting to consider a lter bank modulation scheme where the FIR lters are such that a large attenuation of the lter frequency responses outside the assigned subchannel bands is achieved and that the conditions (9.6) are satised. These objectives are achieved by OFDM systems that are usually known as discrete wavelet multitone modulation (DWMT). In a DWMT system, the elements of the input vector ak are realvalued symbols and the impulse response of the transmit and receive lter banks are also real-valued.
783
i D 0; : : : ; 2M
(9.81)
also let P .i / .z/, i D 0; 1; : : : ; 2M 1, be the transfer functions of the polyphase components of P0 .z/. The amplitude characteristics of the lters are illustrated in Figure 9.23. The basic idea of DWMT consists in combining pairs of these 2M lters so that they are used with M real input signals. We assume that the impulse response of the prototype lter f pn [0]g is real-valued. There2 f T fore the function P0 e j 2M is symmetric with respect to f D 0. Ideally, the prototype lter is a lowpass lter with bandwidth equal to 1=.2T / Hz. From (9.31) we nd that the in impulse response of the i-th lter is given by pn [i] D pn [0] W2M . We consider now a shifted version of 1=.2T / Hz of the original set of 2M frequency responses, obtained by 1=2 the change of variable z ! zW 2M . We dene
1 iC 2 0 i 2M 1 (9.82) Q i .z/ D P0 zW 2M The amplitude characteristics of the shifted frequency responses are illustrated in Figure 9.24. As the coefcients of P0 .z/ are real, the property Q 2M 1 i .z/ D Q i .z /
Figure 9.23. Amplitude of the frequency responses of the lters of an OFDM system with 2M subchannels.
Figure 9.24. Amplitude of the frequency responses of the lters shifted in frequency.
784
(9.83)
0i M
1 i .z/
1 1
Vi .z/ D
P0 z W 2M
1 iC 2
D i Q 2M
0i M
D i Q i .z /
and we dene the transfer functions of a new lter bank with M transmit lters, real input symbol sequences, and modulation rate equal to 2=T , as Hi .z/ D i Ui .z/ C i Vi .z/ 0i M 1 (9.87)
In the previous equations i and i are constants with absolute value equal to one. The amplitude of the frequency response of the lter Hi .z/ is illustrated in Figure 9.25. We note that Hi .z/ has a frequency response with positive frequency content due to U i .z/, and negative frequency content due to Vi .z/. We assume that the original prototype lter P0 .z/ is an FIR lter with length M and transfer function given by P0 .z/ D
M 1 X
nD0
pn .0/ z
The M lters dened by (9.87) are also FIR lters of length dened as Hi .z/ D
M 1 X
nD0
h n [i] z
0i M
(9.89)
Because the coefcients of P0 .z/ are real-valued, the coefcients of U i .z/ are obtained as the complex conjugate of the coefcients of Vi .z/. Consequently in (9.87) the coefcients h n [i], i D 0; : : : ; M 1, are real-valued.
785
We assume, moreover, that the prototype lter P0 .z/ is a linear-phase lter and that the relation p M 1 n [0] D pn [0] holds. Therefore we get 1 D z . M 1/ P0 .z/ (9.90) P0 z The frequency response of the lter can be expressed as P0 e j 2M
2 f T
De
j2 f
M 1 T 2 2M
PR e j 2M
2 f T
(9.91)
2 f T where PR e j 2M is a real-valued function. We choose the values of the constants i so that Ui .z/ and Vi .z/ have the same linear phase as P0 .z/; observing that fT j2 2M
Ui e we let
D i W2M
1 iC 2
M 1 2
j2 f
M 1 T 2 2M
PR e
fT j2 2M
i C1=2 2M
(9.92)
1 iC 2
M 1 2
(9.93)
De
j2 f
M 1 T 2 2M
PR e
fT j2 2M
i C1=2 2M
(9.94)
De
M 1 T j2 f 2 2M
PR e
f T i C1=2 j2 2M C 2M
and the functions Ui .z/ and Vi .z/ indeed exhibit the same linear phase as P0 .z/. Because Ui .z/ and Vi .z/ have a linear phase, analogously to (9.90), the following relations hold: 1 Ui D z . M 1/ Ui .z/ z (9.95) 1 . M 1/ Vi Vi .z/ Dz z Moreover, we assume that the receive lters are matched, that is gn [i] D h M n [i] D h M n [i], 0 i M 1. Hence the transfer functions of the receive lters are given by 1 D z M H j .z 1 / 0 j M 1 (9.96) G j .z/ D z M H j z From (9.87) we get G j .z/ D z
1
[i U j .z/ C i V j .z/]
0 j M
(9.97)
786
H3 H2 H1 H0
1 2
1111 0000 1111 0000 1111 0000 1111 0000 1111 0000 1111 0000 1111 0000 1111 0000
l =0
0 (a)
1111 0000 1111 0000 1111 0000 1111 0000 1111 0000 1111 0000 1111 0000 1111 0000
l =0
1 2
T 2M
l =3
l =0
l =1
l =2
l =3
1 4
2 4
3 4
fT 2
l =1
l =2
l =3 (b)
l =0
l =1
Figure 9.26. (a) Amplitude of the lter frequency responses for M D 4; (b) spectral components of ICI evaluated from the i-th input to the j-th output, for j D 1 and i D 2, after downsampling (see (9.98)).
787
'
X 1 1 M 1 ` z M WM M `D0 1 ` C z M WM
1
1 1 ` ` jC1 U j z M WM U jC1 z M WM j
where we have used the observation that in the case of ideal lters the functions U j1 and V j2 do not overlap in frequency. Therefore we assume that their product is negligible. From the denition (9.85) of the function V j .z/ and from (9.82) we get 1 1 `C jC1 ` D Q jC1 z M WM V j z M WM j V jC1
1 zM
WM
`C jC1
jC1
Qj
1 zM
WM
(9.99)
Substituting the previous equations in (9.98), observing (9.85) and (9.86), and using the ` periodicity of WM , we obtain X 1 1 M 1 ` z M WM [G j .z/H jC1 .z/] # M D M `D0 C j WM jC1 j jC1
1
. jC1/
(9.100) The condition to suppress the ICI in the j-th subchannel due to the signal transmitted on the . j C 1/-th subchannel is therefore given by jC1 j jC1 C j WM j jC1 j jC1 Substitution of (9.93) in (9.101) yields the condition jC1 D j j jC1 (9.102)
. jC1/
D0
(9.101)
Analogously, for the suppression of the ICI in the j-th subchannel due to the signal transmitted on the . j 1/-th channel, the condition j 1 D j 1 is found. We note j j that, setting j D e j' j , we nd that the condition for the approximate suppression of the ICI can be expressed as 'j D 'j
1
(9.103)
Equation (9.103) sets a constraint on the sequence of the phases of the constants j , 0 j M 1; to dene the whole sequence it is necessary to determine the phase of 0 . From (9.87) and (9.97), we observe that G j .z/ H j .z/ D z
1
(9.104)
788
where the products U j .z/ V j .z/ are negligible except for j D 0 and j D M 1. In these 2 f T cases, to avoid that the function H0 e j 2M is distorted at frequencies near zero and 2 f T that the function HM 1 e j 2M is distorted at frequencies near M=T , it must be 2 C 2 D 0 j j Therefore we choose
4 4 a0 D aM 1
j D 0; M
(9.105)
(9.106)
For example, a sequence f' j gM 1 that satises both the (9.103) and the further condition jD0 (9.106) is given by ' j D . 1/ j 4 j D 0; 1; : : : ; M 1 (9.107)
Moreover, from (9.104), the condition for the absence of ISI is expressed as [z
1
j D 0; 1; : : : ; M
(9.108)
Summarizing, for the design of the system we may start from a prototype lter P0 .z/ that approximates a square root raised cosine lter with Nyquist frequency 1=.2T /. This leads to verifying the condition (9.108). The M subchannel lters are obtained using (9.84), (9.85), and (9.87), where i is dened in (9.93) and the phase of i is given in (9.107). An efcient implementation of the transmit lter bank is illustrated in Figure 9.27, where P .i / .z/ are the 2M polyphase components of the prototype lter P0 .z/, and d i D i i , i D 0; 1; : : : ; M 1. An efcient implementation of the receive lter bank is illustrated in Figure 9.28.
W2M
1 iC 2
M 1 2
i D 0; : : : ; M
(9.110)
789
ak [0] T 2 ak [1] T 2
T 2M d0
P (0)(-z
sn
2M
)
z -1 W2M
-1/2
T 2M
T 2M d1
P (1) (-z
2M
)
z -1 W2M
-1/2
ak [M -1] T 2
T 2M dM -1
IDFT
* dM -1
d 1* P d 0*
( 2M -1)
z -1 W2M
-1/2
(-z
2M
Figure 9.27. OFDM system with approximate suppression of the ICI: transmit lter bank.
while JM denotes the M M matrix that has elements equal to one on the antidiagonal and all other elements equal to zero: 3 2 0 0 0 1 6 0 0 1 0 7 7 6 (9.111) JM D 6 : : : : 7 : : : 5 4 : : : : : 1 0 0 0 We assume that the parameters and M that determine the length M of the prototype lter are even numbers. The element tin of matrix T is then given by
n.2 ni 2 2 tin D W2M W2M di C W2M W2M M 1 i / di 1 M 1 i iC n C . 1/ D 2 cos M 2 2 4 n n
(9.112) 1
D 2.cin cos O
sin sin O
i/
0i M
1 0 n 2M
790
Figure 9.28. OFDM system with approximate suppression of the ICI: receive lter bank.
where
cin D cos O
(9.113)
Therefore using the matrix TT we obtain an equivalent block diagram to that of Figure 9.27, as illustrated in Figure 9.29. We give the following denitions: 1. C is the matrix of the M-point discrete cosine transform (DCT), whose element in the position i; n is given by "r # 1 1 2 cos iC nC [C]i;n D M M 2 2 (9.114) i D 0; : : : ; M 1 n D 0; : : : ; M 1
791
2. S is the matrix of the M-point discrete sine transform (DST), whose element in the position i; n is given by "r # 1 1 2 iC nC sin (9.115) [S]i;n D M M 2 2 3.
c
and
(9.116)
and (9.117)
1 0 : : : 0
0 1 : : : 0
0 0 : : : 0
. 1/M
0 0 : : :
792
c .C C c .C
M S/ M S/
2 2
M S/
A1 D
p M
We consider the polyphase representation of the prototype lter P0 .z/ by the 2M polyphase components P .`/ .z/, ` D 0; 1; : : : ; 2M 1, each having =2 coefcients, and dene p0 .z/ D diagfP .0/ . z/; : : : ; P .M
1/
. z/g (9.121)
1/
. z/g
Using (9.119) and (9.120), the vector of the transfer functions of the transmit lters hT .z/ can be expressed as #" T # " A0 p0 .z 2M / 0M 1 1 T .2M 1/ H M H e ; z e h .z/ D z T z z 0M p1 .z 2M / A1 (9.122) Comparing (9.122) with (9.17), we nd that the matrix of the transfer functions of the polyphase components of the transmit lters is given by " T # A0 1 2 2 E.z/ D [z p0 .z /; p1 .z /] (9.123) T A1 We recall from (9.24) that we get a system without ISI and ICI if E H .1=z /E.z/ D I. We consider the product 3 2 1 " T # z p 2 A0 7 6 0 z 1 [A0 ; A1 ] 6 E.z/ E H D [z 1 p0 .z 2 /; p1 .z 2 /] 7 5 4 T z 1 A1 p 2 1 z (9.124) 1 1 T T D p0 .z 2 /A0 A0 p 2 C z 1 p0 .z 2 /A0 A1 p 2 0 1 z z 1 1 T T zp 1 .z 2 /A1 A0 p 2 C p1 .z 2 /A1 A1 p 2 0 1 z z Using denition (9.120) of matrices A0 and A1 we obtain 1 1 1 H 2 2 E.z/ E D 2M p0 .z / p0 2 C p1 .z / p1 2 z z z (9.125) 1 1 2 2 p1 .z /JM p1 2 2M. 1/ 2 p0 .z / JM p0 2 z z
9. Bibliography
793
We recall that each polyphase component P .`/ .z/ has length =2. Moreover, from (9.90), the relation pn [0] D p M 1 n [0] implies the following constraints on the polyphase components: 1 .`/ 2 1 P .2M 1 `/ ` D 0; : : : ; 2M 1 (9.126) P .z/ D z z From property (9.126), we nd that the diagonal matrices p0 .z/ and p 1 .z/ satisfy the relation 1 2 1 . 1/ 2 1 J JM (9.127) p1 .z/ D z M p0 z Then, using (9.127), we nd that the second term in the (9.125) vanishes. Therefore we obtain 1 1 1 D 2M p0 .z 2 / p 2 C p1 .z 2 / p 2 (9.128) E.z/ E H 0 1 z z z Recalling that for two matrices whose product is the identity matrix the commutative property holds, we get E H .1=z / E.z/ D E.z/ E H .1=z / D I if and only if 1 1 1 I (9.129) p0 .z/ p 0 C p1 .z/ p 1 D z z 2M Using (9.129), we nd the conditions on the polyphase components of the prototype lter for perfect suppression of ISI and ICI, given by 1 1 1 .`/ .`/ .MC`/ P 0`M 1 P .z/ C P P .MC`/ .z/ D z z 2M (9.130) The conditions (9.130) can be used for the design of lters for DWMT systems. An efcient lter bank implementation is obtained by the DCT [1].
Bibliography
[1] P. P. Vaidyanathan, Multirate systems and lter banks. Englewood Cliffs, NJ: PrenticeHall, 1993. [2] M. G. Bellanger, G. Bonnerot, and M. Coudreuse, Digital ltering by polyphase network: application to sample-rate alteration and lter banks, IEEE Trans. on Acoustics, Speech and Signal Processing, vol. ASSP-24, pp. 109114, Apr. 1976. [3] G. Cherubini, E. Eleftheriou, S. Olcer, and J. M. Ciof, Filter bank modulation techniques for very high-speed digital subscriber lines, IEEE Communications Magazine, vol. 38, pp. 98104, May 2000.
794
R [4] G. Cherubini, E. Eleftheriou, and S. Olcer, Filtered multitone modulation for veryhigh-speed digital subscriber lines, IEEE Journal on Selected Areas in Communications, June 2002. [5] S. B. Weinstein and P. M. Ebert, Data transmission by frequency-division multiplexing using the discrete Fourier transform, IEEE Trans. on Communications, vol. 19, pp. 628634, Oct. 1971. [6] J. A. C. Bingham, Multicarrier modulation for data transmission: an idea whose time has come, IEEE Communications Magazine, vol. 28, pp. 514, May 1990. [7] H. Sari, G. Karam, and I. Jeanclaude, Transmission techniques for digital terrestrial TV broadcasting, IEEE Communications Magazine, vol. 33, pp. 100109, Feb. 1995. [8] J. S. Chow, J. C. Tu, and J. M. Ciof, A discrete multitone transceiver system for HDSL applications, IEEE Journal on Selected Areas in Communications, vol. 9, pp. 895908, Aug. 1991. [9] S. D. Sandberg and M. A. Tzannes, Overlapped discrete multitone modulation for high speed copper wire communications, IEEE Journal on Selected Areas in Communications, vol. 13, pp. 15711585, Dec. 1995. [10] P. J. W. Melsa, R. C. Younce, and C. E. Rohrs, Impulse response shortening for discrete multitone transceivers, IEEE Trans. on Communications, vol. 44, pp. 1662 1672, Dec. 1996. [11] R. Baldemair and P. Frenger, A time-domain equalizer minimizing intersymbol and intercarrier interference in DMT systems, in Proc. GLOBECOM 01, San Antonio, TX, Nov. 2001. [12] T. Pollet and M. Peeters, Synchronization with DMT modulation, IEEE Communications Magazine, vol. 37, pp. 8086, Apr. 1999. [13] J. M. Ciof, G. P. Dudevoir, M. V. Eyuboglu, and G. D. Forney, Jr., MMSE decisionfeedback equalizers and coding. Part I and Part II, IEEE Trans. on Communications, vol. 43, pp. 25822604, Oct. 1995.
Chapter 10
The term spread spectrum systems [1, 2, 3, 4, 5, 6, 7] was coined to indicate communication systems in which the bandwidth of the signal obtained by a standard modulation method (see Chapter 6) is spread by a certain factor before transmission over the channel, and then despread, by the same factor, at the receiver. The operations of spreading and despreading are the inverse of each other, i.e. for an ideal and noiseless channel the received signal after despreading is equivalent to the transmitted signal before spreading. For transmission over an ideal AWGN channel these operations do not offer therefore any improvement in performance with respect to a system that does not use spread spectrum. However, the practical applications of spread spectrum systems are numerous, for example, in multipleaccess systems, narrowband interference rejection, and transmission over channels with fading (see Section 10.2).
10.1
We consider the two most common spread spectrum techniques: direct sequence (DS) and frequency hopping (FH).
10.1.1
The baseband equivalent model of a DS system is illustrated in Figure 10.1. We consider the possibility that U users in a multiple-access system simultaneously transmit, using the same frequency band, by code division multiple access (CDMA) (see Section 6.13.2). .u/ The sequence of information bits fb` g of user u undergoes the following transformations.
.u/ 1) Bit-mapper. From the sequence of information bits, a sequence of i.i.d. symbols fak g with statistical power Ma is produced. The symbols assume values in an M-ary constellation, using one of the maps described in Chapter 6; typically, in mobile radio systems, BPSK or QPSK modulation is used. Let T be the symbol period.
2) Spreading. We indicate by the integer N S F the spreading factor, and by Tchi p the chip period. These two parameters are related to the symbol period T by the relation Tchi p D T NSF (10.1)
796
Figure 10.1. Baseband equivalent model of a DS system: (a) transmitter, (b) multiuser channel.
.u/ .u/ NSF 1, u 2 f1; : : : ; U g. Moreover, cCh;m 2 f 1; 1g, and jcCh;m j D 1. The Walsh Hadamard sequences are orthogonal, that is ( SF 1 if u 1 D u 2 1 NX 1 .u 1 / .u 2 / c c D (10.2) NSF mD0 Ch;m Ch;m 0 if u 1 6D u 2 .u/ We assume that the sequences of the channelization code are periodic, that is cCh;m D .u/ We now introduce the user code fcm g, also called signature sequence or spreading sequence, that we initially assume to be equal to the channelization code, .u/ .u/ cm D cCh;m .u/ cCh;m mod N .
SF
We recall from Appendix 6.D the denition of the WalshHadamard sequences of length .u/ N S F . Here we refer to these sequences as channelization code fcCh;m g, m D 0; 1; : : : ;
(10.3)
.u/ Consequently, fcm g is also a periodic sequence of period NSF . .u/ The operation of spreading consists of associating with each symbol ak a sequence of .u/ NSF symbols of period Tchi p , that is obtained as follows. First, each symbol ak is repeated NSF times with period Tchi p : as illustrated in Figure 10.2, this operation is equivalent to .u/ upsampling fak g, so that .NSF 1/ zeros are inserted between two consecutive symbols,
797
a (u) k T
NSF holder
(u) am
(u) dm
Tchip cm
(u)
(a)
a (u) k T
g (u)
sp
(u) dm
Tchip (b)
and using a holder of NSF values. The obtained sequence is then multiplied by the user code. Formally we have
.u/ am D a k N .u/ .u/ N .u/ .u/ dm D am cm
1 (10.4)
i D 0; : : : ; NSF
(10.5)
the correlation of Figure 10.2a can be substituted by an interpolation with the interpolator .u/ lter gsp , as illustrated in Figure 10.2b. .u/ Recalling that jcm j D 1, from (10.4) we get Md D Ma (10.6)
3) Pulse-shaping. Let h Tx be the modulation pulse, typically a square root raised cosine function or rectangular window. The baseband equivalent of the transmitted signal of user u is expressed as s .u/ .t/ D A.u/
C1 X mD 1 .u/ dm h Tx .t
m Tchi p /
(10.7)
where A.u/ accounts for the transmit signal power. In fact, if E h is the energy of h Tx and .u/ fdm g is assumed i.i.d., the average statistical power of s .u/ .t/ is given by (see (1.399)) N Ms .u/ D .A.u/ /2 Md Eh Tchi p (10.8)
798
Using (10.4), an alternative expression for (10.7) is given by s .u/ .t/ D A.u/
C1 X kD 1 .u/ ak NX 1 SF `D0 .u/ c`Ck NSF h Tx .t
.` C k NSF / Tchi p /
(10.9)
.u/ In the scheme of Figure 10.1a we note that, if condition (10.3) holds, then gsp is invariant .u/ with respect to the symbol period, and the two lters gsp and h Tx can be combined into one lter (see also (10.9))
NX 1 SF `D0
.u/ c` h Tx .t
` Tchi p /
(10.10)
C1 X kD 1
.u/ ak h .u/ .t T
kT /
(10.11)
As shown in Figure 10.3, the equivalent scheme to the cascade of spreader and pulse-shaping lter is still a QAM modulator. The peculiarity is that the lter h .u/ has a bandwidth much T larger than the Nyquist frequency 1=.2T /. Therefore a DS system can be interpreted as a .u/ QAM system either with input symbols fdm g and transmit pulse h Tx , or with input symbols .u/ .u/ fak g and pulse h T ; later both interpretations will be used. 4) Transmission channel. Modeling the transmission channel as a lter having impulse .u/ response gC , the output signal is given by
.u/ .u/ sC .t/ D .s .u/ gC /.t/
(10.12)
The possibility that many users transmit simultaneously over the same frequency band leads to a total signal sC .t/ D
U X uD1 .1/ We assume we are interested in reconstructing the message fak g of user u D 1, identied .u/ as desired user. If Ms .u/ is the statistical power of sC , the following signal-to-interference C ratios (SIRs) dene the relative powers of the user signals: Ms .1/ 0i.u/ D C u D 2; 3; : : : ; U (10.14) Ms .u/
C
.u/ sC .t/
(10.13)
a (u) k T
h (u)
T
(u)
s (u)(t)
799
5) Noise. In Figure 10.1b the term wC includes both the noise of the receiver and possible additional interference, such as the interference due to signals coming from other cells in a wireless system; wC is modeled as white noise with PSD equal to N0 . Two signal-to-noise ratios are of interest. To measure the performance of the system in terms of Pbit , it is convenient to refer to the signal-to-noise ratio dened in Chapter 6 for passband transmission, 0s D Ms .1/
C
N0 =T
(10.15)
.1/ We recall that for an uncoded sequence of symbols fak g, the following relation holds:
0s Eb D N0 log2 M
(10.16)
However, there are cases, as for example in the evaluation of the performance of the channel impulse response estimation algorithm, when it is useful to measure the power of the noise over the whole transmission bandwidth. Hence we dene the ratio 0c D Ms .1/
C
N0 =Tchip
0s NSF
(10.17)
6) Receiver. The receiver structure varies according to the channel model and number of users. Deferring until Section 10.3 the analysis of more complicated system congurations, .u/ here we limit ourselves to considering the case of an ideal AWGN channel with gC .t/ D .t/ and synchronous users. The latter assumption implies that the transmitters of the various users are synchronized and transmit at the same instant. For an ideal AWGN channel this means that at the receiver the optimum timing phase of signals of different users is the same. With these assumptions, we verify that the optimum receiver is simply given by the .1/ matched lter to h T .t/. According to the analog or discrete-time implementation of the matched lters, we get the schemes of Figure 10.4 or Figure 10.5, respectively; note that in Figure 10.5 the receiver front-end comprises an anti-aliasing lter followed by a sampler with sampling period Tc D Tchip =2. Let t0 be the optimum timing phase at the matched lter output (see Section 14.7). For an ideal AWGN channel it results in rC .t/ D
U X uD1
U X uD1
A.u/
ai.u/
NX 1 SF `D0
.u/ c` h Tx .t
In the presence only of the desired user, that is for U D 1, it is clear that in the absence of ISI the structure with the matched lter to h .1/ .t/ is optimum. We verify that the presence T of other users is cancelled at the receiver, given that the various user codes are orthogonal.
800
g (t) =h (1)*(t)
T
(1) M
r (t) C
(1) M
yk T (a)
^ (1) ak
g (t) =h * (t)
M Tx
r (t) C
yk T
^ (1) ak
Figure 10.4. Optimum receiver with analog lters for a DS-CDMA system with ideal AWGN channel and synchronous users. Two equivalent structures: (a) overall matched lter, (b) matched lter to hTx and despreading correlator.
r (t) C
t 0 +nTc g r
AA (t) AA
g (t) =h (1)*(t)
T (1) g M
(1) M
Tc =
2NSF
yk T a)
^ (1) ak T
g Tchip 2 g Tchip 2
xm Tchip
g (1)
ds
N SF
yk b)
xm Tchip
(1)* cm
yk c)
despreading
Figure 10.5. Optimum receiver with discrete-time lters for a DS-CDMA system with ideal AWGN channel and synchronous users. Three equivalent structures: (a) overall matched lter, (b) matched lter of hTx and despreading lter, (c) matched lter to hTx and despreading correlator.
801
We assume that the overall analog impulse response of the system is a Nyquist pulse, hence
.u/ .h Tx gC g A A g M /.t/jtDt0 C j Tchip D .h T x g M /.t/jtDt0 C j Tchi p D E h j
(10.19)
We note that, if t0 is the instant at which the peak of the overall pulse at the output of g M 0 0 is observed, then t0 in Figure 10.5 is given by t0 D t0 tg M , where tg M is the duration of g M . Moreover, from (10.19), we get that the noise at the output of g M , sampled with sampling rate 1=Tchip , wm D .wC g M /.t/jtDt0 Cm Tchip Q (10.20)
is an i.i.d. sequence with variance N0 E h . Hence, from (10.9) and (10.19), the signal at the output of g M , sampled with sampling rate 1=Tchip , has the following expression: xm D E h
U X uD1
A.u/
C1 X iD 1
ai.u/
NX 1 SF `D0
` i NSF
C wm Q
(10.21)
7) Despreading. We now correlate the sequence of samples fxm g, suitably synchronized, with the code sequence of the desired user, and we form the signal yk D
NX 1 SF jD0 .u/ As usual, introducing the lter gds given by .u/ gds .i Tchip / D c.1/ NSF 1 i
.1/
(10.23)
i D 0; 1; : : : ; NSF
(10.24)
the correlation (10.23) is implemented through the lter (10.24), followed by a downsampler, as illustrated in Figure 10.5b. Substitution of (10.22) in (10.23) yields yk D NSF E h where in general rc.u 1 / c.u 2 / .n D / D nD D 1 NSF .NSF jn D j
NSF Xjn D j 1 jD0
1 2 c jCk NSF Cn D c jCk NSF
U X uD1
(10.25)
.u /
.u /
(10.26)
1/; : : : ; 1; 0; 1; : : : ; N SF
802
In the considered case, from (10.3) and (10.2) we get rc.u 1 / c.u 2 / .0/ D u 1 (10.25) simply becomes
.1/ yk D NSF E h A.1/ ak C wk .1/ where N S F E h is the energy of the pulse associated with ak (see (10.10)). In (10.27) the noise is given by NX 1 SF jD0
u2 .
Therefore
(10.27)
wk D
(10.28)
(10.29)
8) Data detector. Using a threshold detector, from (10.27) the signal-to-noise ratio at the decision point is given by (see (7.106)) D dmin 2 I 2 D NSF E h .A.1/ /2 .NSF E h A.1/ /2 D NSF N0 E h =2 N0 =2 (10.30)
On the other hand, from (10.8) and (10.15) we get 0s D .A.1/ /2 Ma E h =Tchip NSF E h .A.1/ /2 Ma Es D D N0 =.NSF Tchip / N0 N0 (10.31)
where E s is the average energy per symbol of the transmitted signal. In other words, the relation between and 0s is optimum, as given by (7.114). Therefore, with regard to user u D 1, at the decision point the system is equivalent to an M-QAM system. However, as observed before, the transmit pulse h .1/ has a bandwidth much larger T than 1=.2T /. 9) Multi-user receiver. The derivation of the optimum receiver carried out for user 1 can be repeated for each user. Therefore we obtain the multiuser receiver of Figure 10.6, composed of a matched lter to the transmit pulse and a despreader bank, where each branch employs a distinct user code. We observe that for the ideal AWGN channel case, spreading the bandwidth of the transmit signal by a factor U allows the simultaneous transmission of U messages using the same frequency band.
803
g (1)
ds
^ (1) ak T ^ (2) ak T
g (t) =h * (t)
M Tx
g (2)
ds
r (t) C
Tchip
(U) g ds
^ (U) ak T
Figure 10.6. Multiuser receiver for a CDMA synchronous system with an ideal AWGN channel.
From the point of view of each mobile station, all U users share the same channel. Therefore, although the channel impulse response depends on the site of the mobile, we have
.u/ gC .t/ D gC .t/
u D 1; : : : ; U
(10.32)
and the residual interference is due to signals originating from adjacent cells in addition to the multipath interference introduced by the channel. In general, interference due to the other users within the same cell is called multi-user interference (MUI) or co-channel interference (CCI). Asynchronous systems. In this case the various user signals are not time-aligned. In a wireless system, this situation typically occurs in the reverse or uplink transmission from the mobile stations to the base station. Because the WalshHadamard codes do not exhibit good cross-correlation properties for lags different from zero, PN scrambling sequences are used (see Appendix 3.A). The user code is then given by
.u/ .u/ cm D cCh;m cscr;m
(10.33)
where fcscr;m g may be the same for all users in a cell. It is necessary now to make an important observation. In some systems the period of fcscr;m g is equal to the length of fcCh;m g, that is NSF , whereas in other systems it is much larger than NSF .1 In the latter case, spreading and despreading operations remain unchanged, .u/ even if they are symbol time varying, as fcm g changes from symbol to symbol; note that consequently the receiver is also symbol time varying.
This observation must not be confused with the distinction between the use of short (of period ' 215 ) or long (of period ' 242 ) PN scrambling sequences, which are employed to identify the base stations or the users and to synchronize the system [8].
804
Asynchronous systems are characterized by codes with low cross-correlation for nonzero lags; however, there is always a residual non-zero correlation among the various user signals. Especially in the presence of multipath channels, the residual correlation is the major cause of interference in the system, which now originates from signals within the cell: for this reason the MUI is usually characterized as intracell MUI.
Synchronization
The despreading operation requires that the receiver is capable of reproducing a user code sequence synchronous with that used for spreading. Therefore the receiver must rst per.u/ form acquisition, that is the code sequence fcm g produced by the local generator must be synchronized with the code sequence of the desired user, so that the error in the time alignment between the two sequences is less than one chip interval. As described in Section 14.7, acquisition of the desired user code sequence is generally obtained by a sequential searching algorithm that, at each step, delays the local code generator by a fraction of a chip, typically half a chip, and determines the correlation between the .u/ signals fxm g and fcm g; the search terminates when the correlation level exceeds a certain threshold value, indicating that the desired time alignment is attained. Following the acquisition process, a tracking algorithm is used to achieve, in the steady state, a time alignment .u/ between the signals fxm g and fcm g that has the desired accuracy; the more commonly used tracking algorithms are the delay-locked loop and the tau-dither loop. The synchronization method also suggests the use of PN sequences as user code sequences. In practice, the chip frequency is limited to values of the order of hundreds of Mchip/s because of the difculty in obtaining an accuracy of the order of a fraction of a nanosecond in the synchronization of the code generator. In turn, this determines the limit in the bandwidth of a DS signal.
10.1.2
The FH spread spectrum technique is typically used for the spreading of M-FSK signals. We consider an M-FSK signal (see Example 6.7.1 on page 486) with carrier frequency complex form, i.e. we consider the analytic signal, as A e j2. f 0 C1 f .t//t , f 0 expressed in P where 1 f .t/ D C1 1 ak wT .t kT /, with fak g sequence of i.i.d. symbols taken from the kD alphabet A D f .M 1/; : : : ; 1; C1; : : : ; M 1g, at the symbol rate 1=T . An FH/M-FSK signal is obtained by multiplying the M-FSK signal by a signal cFH .t/ given by c F H .t/ D
C1 X iD 1
i Thop /
(10.34)
where f f 0;i g is a pseudorandom sequence that determines shifts in frequency of the FH/MFSK signal, f'0;i g is a sequence of random phases associated with the sequence of frequency shifts, and wThop is a rectangular window of duration equal to a hop interval Thop . In an FH/M-FSK system, the transmitted signal is then given by s.t/ D Re[cFH .t/ e j2. f 0 C1 f .t//t ] (10.35)
805
In practice, the signal cFH .t/ is not generated at the transmitter; the transmitted signal s.t/ is obtained by applying the sequence of pseudorandom frequency shifts f f 0;i g directly to the frequency synthesizer that generates the carrier at frequency f 0 . With reference to the implementation illustrated in Figure 10.7, segments of L consecutive chips from a PN sequence, not necessarily disjoint, are applied to a frequency synthesizer that makes the carrier frequency hop over a set of 2 L frequencies. As the band over which the synthesizer must operate is large, it is difcult to maintain the carrier phase coherent between two consecutive hops [9]; if the synthesizer is not equipped with any device to maintain a coherent phase, it is necessary to include a random phase '0;i as in the expression (10.34). In a time interval that is long with respect to Thop , the bandwidth of the signal s.t/, BSS , can be in practice of the order of several GHz. However, in a short time interval during which no frequency hopping occurs, the bandwidth of an FH/M-FSK signal is the same as the bandwidth of the M-FSK signal that carries the information, usually much lower than BSS . Despreading, in this case also called dehopping, is ideally carried out by multiplying the received signal r.t/ by a signal cFH .t/ equal to that used for spreading, apart from O the sequence of random phases associated with the frequency shifts. For non-coherent demodulation, the sequence of random phases can be modelled as a sequence of i.i.d. random variables with uniform probability density in [0; 2 /. The operation of despreading yields the signal x.t/, given by the sum of the M-FSK signal, the noise and possibly interference. The signal x.t/ is then ltered by a lowpass lter and presented to the input of the receive section comprising a non-coherent demodulator for M-FSK signals. As in the case of DS systems, the receiver must perform acquisition and tracking of the FH signal, so that the waveform generated by the synthesizer for dehopping reproduces as accurately as possible the signal cFH .t/.
806
Classication of FH systems
FH systems are traditionally classied according to the relation between Thop and T . Fast frequency-hopped (FFH) systems are characterized by one or more frequency hops per symbol interval, that is T D N Thop , N integer, and slow frequency-hopped (SFH) systems are characterized by the transmission of several symbols per hop interval, that is Thop D N T . Moreover, a chip frequency Fchip is dened also for FH systems, and is given by the largest value among Fhop D 1=Thop and F D 1=T . Therefore the chip frequency Fchip corresponds to the highest among the clock frequencies used by the system. The frequency spacing between the tones of an FH/M-FSK signal is related to the chip frequency and is therefore determined differently for FFH and SFH systems. SFH systems. For SFH systems, Fchip D F, and the spacing between FH/M-FSK tones is equal to the spacing between the M-FSK tones themselves. In a system that uses a non-coherent receiver for M-FSK signals, orthogonality of tones corresponding to M-FSK symbols is obtained if the frequency spacing is an integer multiple of 1=T . Assuming the minimum spacing is equal to F, the bandwidth BSS of an FH/M-FSK signal is partitioned into N f D BSS =F D BSS =Fchip sub-bands with equally spaced center frequencies; in the most commonly used FH scheme the N f tones are grouped into Nb D N f =M adjacent bands without overlap in frequency, each one having a bandwidth equal to M F D M Fchip , as illustrated in Figure 10.8. Assuming M-FSK modulation symmetric around the carrier frequency, the center frequencies of the Nb D 2 L bands represent the set of carrier frequencies generated by the synthesizer, each associated with an L-uple of binary symbols. According to this scheme, each of the N f tones of the FH/M-FSK signal corresponds to a unique combination of carrier frequency and M-FSK symbol.
BSS MF MF MF MF MF
Nf 3 Nf 2 Nf 1 Nf =4Nb Nb
frequency
Figure 10.8. Frequency distribution for an FH/4-FSK system with bands non-overlapping in frequency; the dashed lines indicate the carrier frequencies.
807
BSS MF MF MF MF
MF MF MF
Nf 3 Nf 2 Nf 1 Nf =4Nb
frequency
Figure 10.9. Frequency distribution for an FH/4-FSK system with bands overlapping in frequency.
In a different scheme, that yields a better protection against an intentional jammer using a sophisticated disturbance strategy, adjacent bands exhibit an overlap in frequency equal to .M 1/Fchip Hz, as illustrated in Figure 10.9. Assuming that the center frequency of each band corresponds to a possible carrier frequency, as all N f tones except .M 1/ are available as center frequencies, the number of carrier frequencies increases from N f =M to N f .M 1/, which for N f M represents an increase by a factor M of the randomness in the choice of the carrier frequency. FFH systems. For FFH systems, where Fchip D Fhop , the spacing between tones of an FH/M-FSK signal is equal to the hop frequency. Therefore the bandwidth of the spread spectrum signal is partitioned into a total of N f D BSS =Fhop D BSS =Fchip sub-bands with equally spaced center frequencies, each corresponding to a unique L-uple of binary symbols. Because there are Fhop =F hops per symbol, the metric used to decide upon the symbol with a non-coherent receiver is suitably obtained by summing Fhop =F components of the received signal.
10.2
The most common applications of spread spectrum systems, that will be discussed in the next sections, may be classied as follows. 1. Multiple access. In alternative to FDMA and TDMA systems, introduced in Section 6.13.2, spread spectrum systems allow the simultaneous transmission of messages by several users over the channel, as discussed in Section 10.1.1.
808
2. Narrowband interference rejection. We consider the DS case. Because interference is introduced in the channel after signal spreading, at the receiver the despreading operation compresses the bandwidth of the desired signal to the original value, and at the same time it expands by the same factor the bandwidth of the interference, thus reducing the level of the interference power spectral density. After demodulation the ratio between the desired signal power and the interference power is therefore larger than that obtained without spreading the signal spectrum. 3. Robustness against fading. Widening the signal bandwidth allows exploitation of the multipath diversity of a radio channel affected by fading. Applying a DS spread spectrum technique, intuitively, has the effect of modifying a channel model that is adequate for transmission of narrowband signals in the presence of at fading or multipath fading with a few rays, to a channel model with many rays. Using a receiver that combines the desired signal from the different propagation rays, the power of the desired signal at the decision point increases. In an FH system, on the other hand, we obtain diversity in the time domain, as the channel changes from one hop interval to the next. The probability that the signal is affected by strong fading during two consecutive hop intervals is usually low. To recover the transmitted message in a hop interval during which strong fading is experienced, error correction codes with very long interleaver and ARQ schemes are used (see Chapter 11).
10.2.1
Anti-jam communications
Narrowband interference We consider the baseband equivalent signals of an M-QAM passband communication system with symbol rate F D 1=T , transmitted signal power equal to Ms , and PSD with minimum bandwidth, i.e. Ps . f / D E s rect. f =F/, where E s F D Ms . We now consider the application of a DS spread spectrum modulation system. Due to spreading, the bandwidth of the transmitted signal is expanded from F to BSS D NSF F. Therefore, for the same transmitted signal power, the PSD of the transmitted signal becomes Ps 0 . f / D .E s =NSF / rect. f =BSS /, where E s =NSF D Ms =BSS . We note that spreading has decreased the amplitude of the PSD by the factor NSF , as illustrated in Figure 10.10. In the band of the spread spectrum signal, in addition to additive white Gaussian noise with PSD N0 , we assume the channel introduces an additive interference signal or jammer with power Mj , uniformly distributed on a bandwidth Bj , with Bj < 1=T . With regard to the operation of despreading, we consider the signals after the multiplication by the user code sequence. The interference signal spectrum is expanded and has a PSD equal to Pj0 . f / D Nj rect. f =BSS /, with Nj D Mj =BSS . The noise, that originally has a uniformly distributed power over all the frequencies, still has PSD equal to N0 , i.e. spreading has not changed the PSD of the noise. At the output of the despreading the desired signal exhibits the original PSD equal to E s rect. f =F/. Modeling the despreader lter as an ideal lowpass lter with bandwidth 1=.2T /, for the signal-to-noise ratio at the decision point the following relation holds: 1 Ma 2 D Es Ms =F D N0 C Nj .N0 C Mj =BSS / (10.36)
809
Figure 10.10. Power spectral density of an M-QAM signal with minimum bandwidth and of a spread spectrum M-QAM signal with spreading factor NSF D 4.
In practice, performance is usually limited by interference and the presence of white noise can be ignored. Therefore, assuming Nj N0 , (10.36) becomes Es 1 Ms =F Ms BSS ' Ma (10.37) D D 2 Nj Mj =BSS Mj F where Ms =Mj is the ratio between the power of the desired signal and the power of the jammer, and BSS =F is the spreading ratio N S F also dened as the processing gain of the system. The above considerations are now dened more precisely in the following case. Sinusoidal interference. We assume that the baseband equivalent received signal is expressed as rC .t/ D s.t/ C j.t/ C wC .t/ (10.38)
where s.t/ is a DS signal given by (10.9) with amplitude A.u/ D 1, wC .t/ is AWGN with spectral density N0 , and the interferer is given by j.t/ D Aj e j' (10.39)
p In (10.39) Aj D Mj is the amplitude of the jammer and ' a random phase with uniform p distribution in [0; 2 /. We also assume a minimum bandwidth transmit pulse, h Tx .t/ D p E h =Tchip sinc.t=Tchip /, hence g M .t/ D h Tx .t/, and G M .0/ D E h Tchip . For the coherent receiver of Figure 10.4, at the detection point the sample at instant kT is given by yk D NSF E h ak C wk C Aj e j' G M .0/
NX 1 SF jD0 Modeling the sequence fck NSF ; ck NSF C1 ; : : : ; ck NSF CNSF 1 g as a sequence of i.i.d. random variables, the variance of the summation in (10.40) is equal to NSF , and the ratio
c NSF jCk
(10.40)
810
We note that in the denominator of (10.42) the ratio Mj =Ms is divided by N S F . Recognizing that Mj =Ms is the ratio between the power of the jammer and the power of the desired signal before the despreading operation, and that Mj =.NSF Ms / is the same ratio after the despreading, we nd that, by analogy with the previous case of narrowband interference, also in the case of a sinusoidal jammer the use of the DS technique reduces the effect of the jammer by a factor equal to the processing gain.
10.2.2
Multiple-access systems
Spread spectrum multiple-access communication systems represent an alternative to TDMA or FDMA systems and are normally referred to as CDMA systems (see Section 6.13.2 and Section 10.1.1). With CDMA, a particular spreading sequence is assigned to each user to access the channel; unlike FDMA, where users transmit simultaneously over nonoverlapping frequency bands, or TDMA, where users transmit over the same band but in disjoint time intervals, users in a CDMA system transmit simultaneously over the same frequency band. Because in CDMA systems correlation receivers are usually employed, it is important that the spreading sequences are characterized by low cross-correlation values. We have already observed that CDMA systems may be classied as synchronous or asynchronous. In the rst case the symbol transition instants of all users are aligned; this allows the use of orthogonal sequences as spreading sequences and consequently the elimination of interference caused by one user signal to another; in the second case the interference caused by multiple access limits the channel capacity, but the system design is simplied. CDMA has received particular interest for applications in wireless communications systems, for example, cellular radio systems, personal communications services (PCS), and wireless local-area networks; this interest is mainly due to performance that spread spectrum systems achieve for the transmission over channels characterized by multipath fading. Other properties make CDMA interesting for application to cellular radio systems, for example the possibility of applying the concept of frequency reuse (see Chapter 17). In cellular radio systems based on FDMA or TDMA, to avoid excessive levels of interference from one cell onto neighboring cells, the frequencies used in one cell are not used in neighboring cells. In other words, the system is designed so that there is a certain spatial separation between cells that use the same frequencies. For CDMA, this spatial separation
811
is not necessary, making it possible, in principle, to reuse all frequencies. Moreover, as CDMA systems tend to be limited by interference, an increase in system capacity is obtained by detecting the speech signal activity. This gain is made possible by the fact that in every telephone conversation each user speaks only for about half the time, while in the silence intervals he does not contribute to instantaneous interference. If several users can be served by the system, on average only half of them are active at a given instant, and the effective capacity can be doubled.
10.2.3
Interference rejection
Besides the above described properties, that are relative to the application in multipleaccess systems, the robustness of spread spectrum systems in the presence of narrowband interferers is key in other applications, for example, in systems where interference is unintentionally generated by other users that transmit over the same channel. We have CCI when a certain number of services are simultaneously offered to users transmitting over the same frequency band. Although in these cases some form of spatial separation among signals interfering with each other is usually provided, for example, by using directional antennas, it is often desirable to use spread spectrum systems for their inherent interference suppression capability. In particular, we consider a scenario in which a frequency band is only partially occupied by a set of narrowband conventional signals: to increase the spectral efciency of the system, a set of spread spectrum signals can simultaneously be transmitted over the same band, thus allowing two sets of users to access the transmission channel. Clearly, this scheme can be implemented only if the mutual interference, which a signal set imposes on the other, remains within tolerable limits.
10.3
Before introducing a structure that is often employed in receivers for DS spread spectrum signals, we make the following considerations on the radio channel model introduced in Section 4.6.
gi .
i TQ /
(10.43)
where for simplicity we have assumed the absence of Doppler spread. Therefore the nonzero gains fgi g are uncorrelated random variables and the delays i D i TQ are multiples of a sufciently small period TQ .
812
Hence, from (10.43), the channel output signal sC is related to the input signal s by sC .t/ D
Nc;1 1 X i D0
gi s.t
i TQ /
(10.44)
Now the number of resolvable or uncorrelated rays in (10.44) is generally less than Nc;1 and is related to the bandwidth of s by the following rule: if s has a bandwidth B, the uncorrelated rays are spaced by a delay of the order of 1=B. Consequently, for a channel with a delay spread r ms and bandwidth B / 1=Tchip , the number of resolvable rays is given by Nc;r es / r ms Tchip (10.45)
Using the notion of channel coherence bandwidth, Bccb / 1=r ms , (10.45) may be rewritten as Nc;r es / B Bccb (10.46)
We now give an example that illustrates the above considerations. Let fgC .nTQ /g be a realization of the channel impulse response with uncorrelated coefcients having a given power delay prole; the innite bandwidth of the channel will be equal to B D 1=.2TQ /. We now lter fgC .nTQ /g with two lters having, respectively, bandwidth B D 0:1=.2TQ / and B D 0:01=.2TQ /, and we compare the three pulse shapes given by the input sequence and the two output sequences. We note that the output obtained in correspondence of the lter with the narrower bandwidth has fewer resolvable rays. In fact, in the limit for B ! 0 the output is modeled as a single random variable. Another way to derive (10.45) is to observe that, for t within an interval of duration 1=B, s does not vary much. Therefore, letting Ncor D equation (10.44) can be written as sC .t/ D where gr es; j '
NX 1 cor i D0 Nc;r es 1 X jD0
Nc;1 Nc;r es
(10.47)
gr es; j s.t
j Ncor TQ /
(10.48)
gi C j Ncor
(10.49)
are the gains of the resolvable rays. The conclusion is that, assuming the symbol period T is given and DS spread spectrum modulation is adopted, the larger, the NSF , the greater the resolution of the radio channel, that is, the channel can be modeled with a larger number of uncorrelated rays, with delays of the order of Tchip .
813
and let g M .t/ D qC .t0 t/ be the corresponding matched lter. In practice, at the output of the lter gAA an estimate of qC with sampling period Tc D Tchip =2 is evaluated,2 which yields the corresponding discrete-time matched lter with sampling period of the input signal equal to Tc and sampling period of the output signal equal to Tchip (see Figure 10.11). If qC is sparse, that is, it has a large support but only a few non-zero coefcients, for the realization of g M we retain only the coefcients of qC with larger amplitude; it is better to set to zero the remaining coefcients because their estimate is usually very noisy (see Appendix 3.A). Figure 10.12a illustrates in detail the receiver of Figure 10.11 for a lter g M with at most NMF coefcients spaced of Tc D Tchip =2. If we now implement the despreader on every branch of the lter g M , we obtain the structure of Figure 10.12b. We observe that typically only 3 or 4 branches are active, that is they have a coefcient g M;i different from zero. Ideally, for an overall channel with Nr es resolvable paths, we assume
Nr es X i D1
qC;i .t
i /
(10.51)
Nr es X jD1
qC; j .t0
j/
(10.52)
j D 1; : : : ; Nr es
(10.53)
the receiver scheme, analogous to that of Figure 10.12b, is illustrated in Figure 10.13.
t 0 +nTc g rAA (t)
AA
g (t) =q * (t t) C 0 M r
AA,n
r (t) C
yk T
^ (1) ak T
Tc =
Tchip 2
To determine the optimum sampling phase t0 , usually r A A is oversampled with a period TQ such that Tc =TQ D 2 or 4 for Tc D Tchip =2; among the 2 or 4 estimates of gC obtained with sampling period Tc , the one with the largest energy is selected (see Observation 8.5 on page 641).
814
AA,2m
Tc =
Tchip
Tc
Tc
M,1
Tc
M,0
M,NSF 2
M,NMF1
xm Tchip
g (1)
ds
yk
^ (1) ak
(a) r
AA,2NSF k
T Tc = chip 2
Tc
Tc
2 Tchip
Tchip
despreader
g (1)
ds
g (1)
ds
T
g
M,0
T g
M,NMF 1
yk
T (b)
^ (1) ak
Figure 10.12. Two receiver structures: (a) chip matched lter with despreader, (b) rake.
To simplify the analysis, we assume that the spreading sequence is a PN sequence with N S F sufciently large, such that the following approximations hold: 1) the autocorrelation of the spreading sequence is a Kronecker delta: and 2) the delays fi g are multiples of Tchip . From (10.51), in the absence of noise the signal r A A is given by
Nr es X nD1 C1 X iD 1
r A A .t/ D
qC;n
ai.1/
NX 1 SF `D0
.` C i NSF / Tchip /
(10.54)
815
despreader
Figure 10.13. Rake receiver for a channel with Nres resolvable paths.
qC;n
C1 X iD 1
ai.1/
NX 1 SF `D0
j n mC T chip
.`Ci NSF /
(10.55)
Correspondingly the despreader output, assuming rc.1/ .n D / D n D and the absence of .1/ noise, yields the signal NSF ak qC; j . The contributions from the various branches are then combined according to the MRC technique (see Section 6.13) to yield the sample ! Nr es X .1/ 2 yk D NSF jqC;n j ak (10.56)
nD1
where E qC D nD1 jqC;n j2 is the energy per chip of the overall channel impulse response. The name rake originates from the structure of the receiver that is similar to a rake with Nr es ngers. In practice, near the rake receiver a correlator estimates the delays, with precision Tchip =2, and the gains of the various channel rays. The rake is initialized with the
3
P Nr es
Instead of using the Dirac delta in (10.51), a similar analysis assumes that 1) gAA .t/ D h . t/, and 2) rh Tx .t/ Tx is a Nyquist pulse. The result is the same as (10.55).
816
coefcients of rays with larger gain. The delays and the coefcients are updated whenever a change in the channel impulse response is observed. However, after the initialization has taken place, on each nger of the rake the estimates of the amplitude and of the delay of the corresponding ray may be rened by using the correlator of the despreader, as indicated by the dotted line in Figure 10.13. We note that if the channel is static, the structure of Figure 10.12a with Tc D Tchip =2 yields a sufcient statistic.
10.4
Interference
For a dispersive channel and in the case of U users, we evaluate the expression of the signal yk at the decision point using the matched lter receiver of Figure 10.11. Similarly to (10.50), we dene
.u/ .u/ qC .t/ D .h Tx gC gAA /.t/
u D 1; : : : ; U
(10.57)
and let
.v/ g .v/ .t/ D qC .t0 M
t/
v D 1; : : : ; U
(10.58)
.u/ be the corresponding matched lter. Moreover, we introduce the correlation between qC .v/ and qC , expressed by .u/ .v/ rq .u/ q .v/ . / D .qC .t/ qC . t//. /
C C
(10.59)
Assuming without loss of generality that the desired user signal has the index u D 1, we .1/ refer to the receiver of Figure 10.11, where we have g M .t/ D g .1/ .t/ D qC .t0 t/, and M xm D
U X uD1
A.u/
C1 X iD 1
ai.u/
NX 1 SF `D0
i NSF / Tchip / C wm Q
(10.60)
A.u/
ai.u/
NX 1 NX 1 SF SF `D0 jD0
(10.61)
iD 1
rq .u/ q .1/ .. j
C C
` C .k
10.4. Interference
817
Introducing the change of index n D ` j and recalling the denition of cross-correlation between two code sequences (10.26), the double equation in ` and j in (10.61) can be written as
1 X
nD .NSF 1/ NX 1 SF nD0
rq .u/ q .1/ .. n C .k
C C
rq .u/ q .1/ .. n C .k
C C
where, to simplify the notation, we have assumed that the user code sequences are periodic of period NSF . The desired term in (10.61) is obtained for u D 1; as r.1/ . n/ D rc.1/ .n/, it has the c following expression: A.1/
C1 X iD 1
ai.1/
NX 1 SF nD .NSF 1/
.NSF
i/ NSF / Tchip /
(10.63)
Consequently, if the code sequences are orthogonal, that is rc.1/ .n/ D n and in the absence of ISI, that is rq .1/ .i NSF Tchip / D i E q .1/
C C
(10.64)
(10.65)
where E q .1/ is the energy per chip of the overall pulse at the output of the lter g A A , then C the desired term (10.63) becomes
.1/ A.1/ NSF E qC ak
(10.66)
which coincides with the case of an ideal AWGN channel (see (10.27)). Note that using the same assumptions we nd the rake receiver behaves as an MRC (see (10.56)). If (10.64) is not veried, as happens in practice, and if rq .1/ .nTchip / 6D n E qC
C
(10.67)
the terms for n 6D 0 in (10.63) give rise to intersymbol interference, in this context also called inter-path interference (IPI). Usually the smaller the NSF , the larger the IPI. We note, however, that if the overall pulse at the output of the CMF is a Nyquist pulse, that is rq .1/ .nTchip / D n E qC
C
(10.68)
then there is no IPI, even if (10.64) is not veried. With reference to (10.62) we observe that, in the multiuser case, if rc.u/ c.1/ .n/ 6D 0 then yk is affected by MUI, whose value increases as the cross-correlation between the pulses .u/ .1/ qC and qC increases.
818
g (1)
M
(1) xm
Tchip
detector
^ (1) ak a)
rC (t) g (U)
M (U) xm
Tchip
detector
^ (U) ak
g (1)
M
(1) xm
Tchip
multiuser detector
^ (1) a
k
rC (t) g (U)
M (U) xm
b) ^ (U) a
k
Tchip
10.5
819
wC (t)
(1) dm
(1)
t +nTc
0
Tchip
h Tx *g (1)
C
AA
g Tc = Tchip 2
CE
~ dm Tchip
g (1)
ds
yk T
^ (1) a kD T
dm
(U)
(U)
Tchip
h Tx *g (U)
C
where fdm g is assumed i.i.d. We distinguish the two following cases: 1) All code sequences are known. This is the case that may occur for downlink transmission .u/ in wireless networks. Then gC .t/ D gC .t/, u D 1; : : : ; U , and we assume dm D
U X uD1 .u/ dm
(10.70)
that is, for the equalizer design, all user signals are considered as desired signals. 2) Only the code sequence of the desired user signal is known. assume
.1/ dm D dm
(10.71)
The other user signals are considered as white noise, with overall PSD Ni , that is added to wC . .1/ From the knowledge of qC and the overall noise PSD, the minimum of the cost function dened in (10.69) is obtained by following the same steps developed in Chapter 8. Obviously, if the level of interference is high, the solution corresponding to (10.71) yields a simple CMF, with low performance whenever the residual interference (MUI and IPI) at the decision point is high. A better structure for single-user detection is obtained by the following approach.
820
wC (t) a (1) k T A
(1)
t +nTc
0
h (1) *g (1) T C
rAA,n Tchip 2
AA
g (1)
SE
2NSF
yk T
^ (1) a kD T
Tc =
a (U)
k
(U)
(U) *g (U) T C
(10.72)
.1/ Note that gSE , that includes also the function of despreading, depends on the code sequence of the desired user. Therefore the length of the code sequence is usually not larger than .1/ NSF , otherwise we would nd a different solution for every symbol period, even if gC is time invariant. Moreover, in this formulation the other user signals are seen as interference, and one of the tasks of gSE is to mitigate the MUI. In an adaptive approach, for example, using the LMS algorithm, the solution is simple to determine and does not require any particular a priori knowledge, except the training .1/ sequence in fak g for initial convergence. On the other hand, using a direct approach we need to identify the autocorrelation of rAA;n and the cross-correlation between rAA;n .1/ and ak D . As usual these correlations are estimated directly or, assuming the messages .u/ fak g, u D 1; : : : ; U , are i.i.d. and independent of each other, we can determine them .u/ using the knowledge of the various pulses fh .u/ g and fgC g, that is the channel impulse T responses and code sequences of all users; for the special case of downlink transmission, the knowledge of the code sequences is sufcient, as the channel is common to all user signals.
10.6
Multiuser detection techniques are essential for achieving near-optimum performance in communication systems where signals conveying the desired information are received in the presence of ambient noise plus multiple-user interference. The leitmotiv of developments in multiuser detection is represented by the reduction in complexity of practical receivers with respect to that of optimal receivers, which is known to increase exponentially with the number of active users and with the delay spread of the channel, while achieving nearoptimum performance. A further element that is being recognized as essential to reap the full benets of interference suppression is the joint application of multiuser detection with other techniques such as spatial-temporal processing and iterative decoding.
821
Here we rst consider the simplest among multiuser receivers. It comprises a bank of U .u/ lters gT , u D 1; : : : ; U , matched to the impulse responses4
.u/ qT .t/ D NX 1 SF `D0 .u/ .u/ c` qC .t
` Tchip /
u D 1; : : : ; U
(10.73)
.u/ where the functions fqC .t/g are dened in (10.57). Decisions taken by threshold detectors on the U output signals, sampled at the symbol rate, yield the detected user symbol sequences. It is useful to introduce this receiver, that we denote as MF, as, substituting the threshold detectors with more sophisticated detection devices, it represents the rst stage of several multiuser receivers, as illustrated in general in Figure 10.17. We introduce the following vector notation. The vector of symbols transmitted by U users in a symbol period T is expressed as .1/ .U ak D [ak ; : : : ; ak / ]T
(10.74)
and the vector that carries the information on the codes and the channel impulse responses of the U users is expressed as
.1/ .U qT .t/ D [qT .t/; : : : ; qT / .t/]T
(10.75)
Joint detectors constitute an important class of multiuser receivers. They effectively mitigate both ISI and MUI, exploiting the knowledge of the vector qT . In particular we consider now block linear receivers: as the name suggests, a block linear receiver is a joint detector that recovers the information contained in a window of K symbol periods. Let
T a D [a0 ; : : : ; aT K .1/ 1] T .1/ .U / T 1; : : : ; aK 1]
D [a0 ; : : : ; a0 ; : : : ; a K
(1) t 0 +m Tchip
.U /
(10.76)
y (1)
(1) g ds k
g (1)
M
Tchip t 0 +m Tchip
(U)
multiuser detector
^ (1) ak
rC (t) g (U)
M
y (U) g ds
(U) k
Tchip
^ (U) ak
We assume that the information on the power of the user signals is included in the impulse responses gC , u D 1; : : : ; U , so that A.u/ D 1, u D 1; : : : ; U .
.u/
822
be the information transmitted by U users and let y be the corresponding vector of K U elements at the MF output. We dene the following correlations:
.u/ .v/ rq.u;v/ .k/ D .qT .t/ qT . t//. /j DkT
(10.77)
with < K , and following the approach in [17, 18, 19, 20], we introduce the K U K U matrix, 3 2 rq .1;1/ .0/ : : : rq .1;U / .0/ : : : rq .1;1/ . / : : : rq .1;U / . / : : : 6 r .2;1/ .0/ : : : r .2;U / .0/ : : : r .2;1/ . / : : : r .2;U / . / : : : 7 q q q 7 6 q 6 : : : : : 7 : : : : : : : : 7 : : : 6 : : : : : 7 : : : 6 6 r .U;1/ .0/ : : : r .U;U / .0/ : : : r .U;1/ . / : : : r .U;U / . / : : : 7 7 6 q q q q 6 : : : : : : :: :: 7 6 : : : : : : : : 7 : : : : : : 7 6 (10.79) TD6 7 6 rq .1;1/ ./ : : : rq .1;U / ./ : : : rq .1;1/ .0/ : : : rq .1;U / .0/ : : : 7 7 6 6 rq .2;1/ ./ : : : rq .2;U / ./ : : : rq .2;1/ .0/ : : : rq .2;U / .0/ : : : 7 7 6 : : : : : 7 : : : 6 : : : : : 7 : : : 6 : : : : : 7 : : : 6 6 rq .U;1/ ./ : : : rq .U;U / ./ : : : rq .U;1/ .0/ : : : rq .U;U / .0/ : : : 7 5 4 : : : : : : :: :: : : : : : : : : : : : : : : Let w be the vector of noise samples at the MF output. It is important to verify that its covariance matrix is N0 T. Then the matrix T is Hermitian and, assuming that it is denite positive, the Cholesky decomposition (2.174) can be applied T D LH L (10.80)
where L H is a lower triangular matrix with positive real elements on the main diagonal. Using (10.76) and (10.79), we nd that the vector y satises the linear relation y D Ta Cw (10.81)
Once the expression (10.81) is obtained, the vector a can be detected by well-known techniques [20]. Applying the zero-forcing criterion, at the decision point we get the vector zDT
1
y
1
(10.82) w
DaCT
Equation (10.82) shows that the zero-forcing criterion completely eliminates both ISI and MUI, but it may enhance the noise. Applying instead the MSE criterion to the signal rC .t/ suitably sampled, leads to the solution (see (2.229)) z D .T C N0 I/
1
(10.83)
823
Both approaches require the inversion of a K U K U Hermitian matrix and therefore a large computational complexity. A scheme that is computationally efcient while maintaining comparable performance is described in [21]. A MMSE method with further reduced complexity operates on single output samples, that is for K D 1. However, the performance is lower because it does not exploit the correlation among the different observations. For the case K D 1, an alternative that yields performance near the optimum ML receiver is represented by a DFE structure (see Section 16.4).
10.7
(10.84)
aiT qT .t
i T / C wC .t/
(10.85)
The log-likelihood associated with (10.84) is [22] 2 Z U X .u/ sC .t/ dt `C D rC .t/ uD1 Dening the matrices Z Qk 1
k2
(10.86)
q .t T
k1 T / qT .t T
k2 T / dt
(10.87)
after several steps, (10.86) can be written as `C D where the branch metric is given by ( "
H MC .k/ D Re ak Q0 ak C K 1 X kD0
MC .k/
(10.88)
X mD1
#) 2Qm ak
m
2yk
(10.89)
We note that the rst two terms within the brackets in (10.89) can be computed off-line. The sequence fO k g that maximizes (10.88) can be obtained using the Viterbi algorithm; a the complexity of this scheme is, however, exceedingly large, because it requires O.42U / branch metric computations per detected symbol, assuming QPSK modulation.
824
Qk D k
(10.91)
where w.D/ is the noisy term with matrix spectral density N0 Q.D/. Assuming that it does not have poles on the unit circle, Q.D/ can be factorized in the form Q.D/ D F H .D
1
/ F.D/
(10.93)
where F.D/ is minimum phase; in particular, F.D/ has the form F.D/ D
X kD0
Fk D k
1 /] 1 ,
where F0 is a lower triangular matrix. Now let .D/ D [F H .D by construction. Applying .D/ to y.D/ in (10.92), we obtain z.D/ D .D/ y.D/ D F.D/ a.D/ C w0 .D/
(10.95)
where the noisy term w0 .D/ is a white Gaussian process. Consequently, in the time domain (10.95) becomes zk D
X mD0
Fm ak
C w0 k
(10.96)
With reference to [23], the expression (10.96) is an extension to the multidimensional case of Forneys MLSD approach. In fact, the log-likelihood can be expressed as the sum of branch metrics dened as M E .k/ D zk
X mD0 2
Fm ak
U X
U X .u/ D z k uD1
(10.97)
We note that, as F0 is a lower triangular matrix, the metric has a causal dependence also with regard to the ordering of the users. For further study on multiuser detection techniques we refer the reader to [24, 25, 26].
10. Bibliography
825
Bibliography
[1] M. K. Simon, J. K. Omura, R. A. Scholtz, and B. K. Levitt, Spread spectrum communications handbook. New York: McGraw-Hill, 1994. [2] R. C. Dixon, Spread spectrum systems. New York: John Wiley & Sons, 3rd ed., 1994. [3] L. B. Milstein and M. K. Simon, Spread spectrum communications, in The Mobile Communications Handbook (J. D. Gibson, ed.), ch. 11, pp. 152165, New York: CRC/IEEE Press, 1996. [4] J. G. Proakis, Digital communications. New York: McGraw-Hill, 3rd ed., 1995. [5] R. Price and P. E. Green, A communication technique for multipath channels, IRE Proceedings, vol. 46, pp. 555570, Mar. 1958. [6] A. J. Viterbi, CDMA: Principles of spread-spectrum communication. Reading, MA: Addison-Wesley, 1995. [7] R. L. Peterson, R. E. Ziemer, and D. E. Borth, Introduction to spread spectrum communications. Englewood Cliffs, NJ: Prentice-Hall, 1995. [8] Wideband CDMA, IEEE Communications Magazine, vol. 36, pp. 4695, Sept. 1998. [9] G. Cherubini and L. B. Milstein, Performance analysis of both hybrid and frequency hopped phasecoherent spreadspectrum system. Part I and Part II, IEEE Trans. on Communications, vol. 37, pp. 600622, June 1989. [10] A. Klein, Data detection algorithms specially designed for the downlink of CDMA mobile radio systems, in Proc. 1997 IEEE Vehicular Technology Conference, Phoenix, USA, pp. 203207, May 47 1997. [11] K. Li and H. Liu, A new blind receiver for downlink DS-CDMA communications, IEEE Communications Letters, vol. 3, pp. 193195, July 1999. [12] S. Werner and J. Lilleberg, Downlink channel decorrelation in CDMA systems with long codes, in Proc. 1999 IEEE Vehicular Technology Conference, Houston, USA, pp. 16141617, May 1620 1999. [13] K. Hooli, M. Latva-aho, and M. Juntti, Multiple access interference suppression with linear chip equalizers in WCDMA downlink receivers, in Proc. 1999 IEEE Global Telecommunications Conference, Rio de Janeiro, Brazil, pp. 467471, Dec. 59 1999. [14] U. Madhow and M. L. Honig, MMSE interference suppression for direct-sequence spread-spectrum CDMA, IEEE Trans. on Communications, vol. 42, pp. 31783188, Dec. 1994.
826
[15] S. L. Miller, An adaptive direct-sequence code-division multiple-access receiver for multiuser interference rejection, IEEE Trans. on Communications, vol. 43, pp. 1746 1755, Feb./Mar./Apr. 1995. [16] P. B. Rapajic and B. S. Vucetic, Adaptive receiver structures for asynchronous CDMA systems, IEEE Journal on Selected Areas in Communications, vol. 12, pp. 685697, May 1994. [17] A. Klein and P. W. Baier, Linear unbiased data estimation in mobile radio systems applying CDMA, IEEE Journal on Selected Areas in Communications, vol. 11, pp. 10581066, Sept. 1993. [18] J. Blanz, A. Klein, M. Nahan, and A. Steil, Performance of a cellular hybrid C/TDMA mobile radio system applying joint detection and coherent receiver antenna diversity, IEEE Journal on Selected Areas in Communications, vol. 12, pp. 568579, May 1994. [19] G. K. Kaleh, Channel equalization for block transmission systems, IEEE Journal on Selected Areas in Communications, vol. 13, pp. 110120, Jan. 1995. [20] A. Klein, G. K. Kaleh, and P. W. Baier, Zero forcing and minimum mean-squareerror equalization for multiuser detection in code-division multiple-access channels, IEEE Trans. on Vehicular Technology, vol. 45, pp. 276287, May 1996. [21] N. Benvenuto and G. Sostrato, Joint detection with low computational complexity for hybrid TD-CDMA systems, IEEE Journal on Selected Areas in Communications, vol. 19, pp. 245253, Jan. 2001. [22] G. E. Bottomley and S. Chennakeshu, Unication of MLSE receivers and extension to time-varying channels, IEEE Trans. on Communications, vol. 46, pp. 464472, Apr. 1998. [23] A. Duel-Hallen, A family of multiuser decision feedback detectors for asynchronous code-division multiple access channels, IEEE Trans. on Communications, vol. 43, pp. 421434, Feb./Mar./Apr. 1995. [24] S. Verd` , Multiuser detection. Cambridge: Cambridge University Press, 1998. u [25] Multiuser detection techniques with application to wired and wireless communications systems I, IEEE Journal on Selected Areas in Communications, vol. 19, Aug. 2001. [26] Multiuser detection techniques with application to wired and wireless communications systems II, IEEE Journal on Selected Areas in Communications, vol. 20, Feb. 2002.
Chapter 11
Channel codes
Forward error correction (FEC) is a widely used technique to achieve reliable data transmission. The redundancy introduced by an encoder for the transmission of data in coded form allows the decoder at the receiver to detect and partially correct errors. An alternative transmission technique, known as automatic repeat query or request (ARQ), consists in detecting the errors (usually by a check-sum transmitted with the data, see page 875) and requesting the retransmission of a data packet whenever it is received with errors. The FEC technique presents two advantages with respect to the ARQ technique. 1. In systems that make use of the ARQ technique the data packets do not necessarily have to be retransmitted until they are received without errors; however, for large values of the error probability, the aggregate trafc of the link is higher. 2. In systems that make use of the FEC technique the receiver does not have to request the retransmission of data packets, thus making possible the use of a simplex link (see Section 6.13); this feature represents a strong point in many applications like TDMA and video satellite links, where a central transmitter broadcasts to receiveonly terminals, which are unable to make a possible retransmission request. The FEC technique is also particularly useful in various satellite communication applications, in which the long round-trip delay of the link would cause serious trafc problems whenever the ARQ technique would be used. We distinguish two broad classes of FEC techniques, each with numerous subclasses, employing block codes or convolutional codes. All error correction techniques add redundancy, in the form of additional bits, to the information bits that must be transmitted. Redundancy makes the correction of errors possible and for the classes of codes considered in this chapter represents the coding overhead. The effectiveness of a coding technique is expressed in terms of the coding gain, G code , given by the difference between the signal-to-noise ratios, in dB, that are required to achieve a certain bit error probability for transmission without and with coding (see Denition 6.2 on page 508). The overhead is expressed in terms of the code rate, Rc , given by the ratio between the number of information bits and the number of code bits that are transmitted. The transmission bit rate is inversely proportional to Rc , and is larger than that necessary for uncoded data. If one of the modulation techniques of Chapter 6 is employed, the modulation rate is also larger. In Chapter 12 methods to
828
transmit coded sequences of symbols without an increase in the modulation rate will be discussed. For further study on the topic of error correcting codes we refer to [1, 2, 3].
11.1
System model
With reference to the model of a transmission system with coding, illustrated in Figure 6.20, we introduce some fundamental parameters. A block code is composed of a set of vectors of given length called code words; the length of a code word is dened as the number of vector elements, indicated by n 0 . The elements of a code word are chosen from an alphabet of q elements: if the alphabet consists of two elements, for example 0 and 1, the code is a binary code and we refer to the elements of each code word as bits; if, on the other hand, the elements of a code word are chosen from an alphabet having q elements .q > 2/, the code is nonbinary. It is interesting to note that if q is a power of two, that is q D 2b , where b is a positive integer, each q-ary element has an equivalent binary representation of b bits and therefore a nonbinary code word of length N can be mapped to a binary code word of length n 0 D bN . There are 2n 0 possible code words in a binary code of length n 0 . From these 2n 0 possible code words, we choose 2k0 words (k0 < n 0 ) to form a code. Thus a block of k0 information bits is mapped to a code word of length n 0 chosen from a set of 2k0 code words; the resulting block code is indicated as .n 0 ; k0 / code and the ratio Rc D k0 =n 0 is the code rate.1 Observation 11.1 The code rate Rc is related to the encoder-modulator rate R I (6.93) by the following relation RI D log M k0 log2 M D Rc 2 n0 I I (11.1)
where M is the number of symbols of the I -dimensional constellation adopted by the bit-mapper. Because the number of bits per unit of time produced by the encoder is larger than that produced by the source, two transmission strategies are possible. Transmission for a given bit rate of the information message. With reference to Figure 6.20, from the relation k0 Tb D n 0 Tcod we obtain 1 1 1 D Tcod Rc Tb
1
(11.2)
(11.3)
In this chapter a block code will sometimes be indicated also by the notation .n; k/.
829
note that the bit rate at the modulator input is increased in the presence of the encoder. For a given modulator with M symbols, that is using the same bit mapper, this implies an increase of the modulation rate given by 1 1 1 1 D log2 M D 0 T Tcod T Rc (11.4)
and therefore an increase of the bandwidth of the transmission channel by a factor 1=Rc . Moreover, for the same transmitted power, from (6.105) in the presence of the encoder the signal-to-noise ratio becomes (11.5) 0 0 D 0 Rc i.e. it decreases by a factor Rc with respect to the case of transmission of an uncoded message. Therefore, for a given information message bit rate 1=Tb , the system operates with a lower 0 0 : consequently the receiver is prone to introduce more bit errors at the decoder input. In spite of this, for a suitable choice of the code, in many cases the decoder O produces a detected message fb` g affected by fewer errors with respect to the case of transmission of an uncoded message. We note that the energy per information bit of the encoded message fcm g is equal to that of fb` g. In fact, for a given bit mapper, L0 D b k0 log2 M D Rc L b n0 (11.6)
Assuming the same transmitted power, from (6.97) and (11.4) we get
0 E sCh D Rc E sCh
(11.7)
Therefore (6.99) yields for the encoded message fcm g an energy per bit of information equal to
0 E sCh
L0 b
E sCh D Eb Lb
(11.8)
Since 0 0 6D 0, a comparison between the performance of the two systems, with and without coding, is made for the same E b =N0 . In this case the coding gain, in dB, is given by 10 .log10 0 log10 0 0 C log10 Rc / (11.9)
Transmission for a given modulation rate. For given transmitted power and given transmission channel bandwidth, 0 remains unchanged in the presence of the encoder. Therefore, there are three possibilities. 1. The bit rate of the information message decreases by a factor Rc and becomes 1 1 0 D T Rc Tb b (11.10)
830
2. The source emits information bits in packets and each packet is followed by additional bits generated by the encoder, forming a code word; the resulting bits are transmitted at the rate 1 1 D (11.11) Tcod Tb 3. A block of m information bits is mapped to a transmitted symbol using a constellation with cardinality M > 2m . In this case transmission occurs without decreasing the bit rate of the information message. In the rst two cases, for the same number of bits of the information message we have an increase in the duration of the transmission by a factor 1=Rc . O For a given bit error probability in the sequence fb` g, we expect that in the presence of coding a smaller 0 is required to achieve a certain error probability as compared to the case of transmission of an uncoded message; this reduction corresponds to the coding gain (see Denition 6.2 on page 508).
11.2
Block codes
We give the following general denition.2 Denition 11.1 The Hamming distance between two vectors v1 and v2 , d H .v1 ; v2 /, is given by the number of elements in which the two vectors differ.
11.2.1 Properties
A binary block code of length n is a subset containing Mc of the 2n possible binary sequences of length n, also called code words. The only requirement on the code words is that they are all of the same length. Denition 11.2 The minimum Hamming distance of a block code, to which we will refer in this chapter H simply as the minimum distance, is denoted by dmin and coincides with the smallest number of positions in which any two code words differ.
H An example of a block code with n D 4; Mc D 4 and dmin D 2 is given by (11.22). For the binary symmetric channel model (6.90), assuming that the binary code word c of length n is transmitted, we observe at the receiver3
zDce
(11.12)
The material presented in Sections 11.2 and 11.3 is largely based on lectures given at the University of California, San Diego, by Professor Jack K. Wolf [4], whom the authors gratefully acknowledge. 3 In Figure 6.20, z is indicated as c. Q
831
where denotes the modulo 2 sum of respective vector components; for example .0111/ .0010/ D .0101/. In (11.12), e is the binary error vector whose generic component is equal to 1 if the channel has introduced an error in the corresponding bit of c, and 0 otherwise. We note that z can assume all the 2n possible combinations of n bits. With reference to Figure 6.20, the function of the decoder consists in associating with each possible value z a code word. A commonly adopted criterion is to associate z with O the code word c that is closest according to the Hamming distance. From this code O word the k0 information bits, which form the sequence fbl g, are recovered by inverse mapping. Interpreting the code words as points in an n-dimensional space where the distance between points is given by the Hamming distance, we obtain the following properties.
H 1. A binary block code with minimum distance dmin can correct all patterns of
tD
jdH
min
1k
(11.13)
1/
3. In a binary erasure channel, the transmitted binary symbols are detected using a ternary alphabet f0; 1; erasureg; a symbol is detected as erasure if the reliability of a binary decision is low. In the absence of errors, a binary block code with minimum H H distance dmin can ll in .dmin 1/ erasures.
H H 4. Seeking a relation among n, Mc , and dmin we nd that, for xed n and odd dmin , Mc 4 is upper bounded by 9 8 = k < j n n n Mc MU B D 2n (11.15) 1C C C C jdH 1k ; min : 1 2 2 H 5. For xed n and dmin , it is always possible to nd a code with Mc words where
Mc D 2n
n n n 1C C C C H 1 2 dmin 1
(11.16)
where dxe denotes the smallest integer greater than or equal to x. We will now consider a procedure for nding such a code.
We recall that the number of binary sequences of length n with m ones is equal to n n! D m!.n m/! m where n! D n.n 1/ 1.
(11.14)
832
Step 1: choose any code word of length n and exclude from future H and all words that differ from it in .dmin 1/ or fewer positions. of words excluded from future choices is n n n H Nc .n; dmin 1/ D 1 C C C C H 1 2 dmin
Step i: choose a word not previously excluded and exclude from future choices all words previously excluded plus the chosen word and those that differ from it in H .dmin 1/ or fewer positions. Continue this procedure until there are no more words available to choose from. At H each step, if still not excluded, at most Nc .n; dmin 1/ additional words are excluded; H therefore after step i, when i code words have been chosen, at most i Nc .n; dmin 1/ n =N .n; d H words have been excluded. Then, if 2 1/ is an integer, we can choose c min at least that number of code words; if it is not an integer, we can choose at least a number of code words equal to the next largest integer. Denition 11.3 A binary code with group structure is a binary block code for which the following conditions are veried: 1. the all zero word is a code word (zero code word); 2. the modulo 2 sum of any two code words is also a code word. Denition 11.4 The weight of any binary vector x, denoted as w.x/, is the number of ones in the vector.
H Property 1 of a group code. The minimum distance of the code dmin is given by H dmin D min w.c/
(11.18)
where c can be any non-zero code word. Proof. The sum of any two distinct code words is a non-zero word. The weight of the resulting word is equal to the number of positions in which the two original words differ. H Because two words at the minimum distance differ in dmin positions, there is a word of H H weight dmin . If there were a non-zero word of weight less than dmin , it would be different H positions. from the zero word in less than dmin Property 2 of a group code. If all code words in a group code are written as rows of an Mc n matrix, then every column is either zero or consists of half zeros and half ones. Proof. An all zero column is possible if all code words have a zero in that column. Suppose in column i there are m 1s and .Mc m/ 0s. Choose one of the words with a 1 in that column and add it to all words that have a 1 in that column, including the word itself: this
833
operation produces m words with a 0 in that column, hence .Mc m/ m. Now we add that word to each word that has a 0 in that column: this produces .Mc m/ words with a 1 in that column, hence .Mc m/ m. Therefore Mc m D m or m D Mc =2. Corollary 11.1 From Property 2 it turns out that the number of code words Mc must be even for a binary group code. Corollary 11.2 Excluding codes of no interest from the transmission point of view, for which all code words have a 0 in a given position, from Property 2 the average weight of a code word is equal to n=2.
where B is an r r matrix with det[B] 6D 0, i.e. the columns of B are linearly independent. A binary parity check code is a code consisting of all binary vectors c that are solutions of the equation Hc D 0 (11.20) The matrix product in (11.20) is computed using the modulo 2 arithmetic. Example 11.2.1 Let the matrix H be given by " HD 1 0 1 1 0 1 0 1 # (11.21)
There are four code words in the binary parity check code corresponding to the matrix H; they are 2 3 2 3 2 3 2 3 0 1 0 1 607 607 617 617 c1 D 6 7 c2 D 6 7 c3 D 6 7 (11.22) c0 D 6 7 405 415 415 405 0 0 1 1 Property 1 of a parity check code. A parity check code is a group code. Proof. The all zero word is always a code word, as H0 D 0 (11.23)
834
Suppose that c1 and c2 are code words; then Hc1 D 0 and Hc2 D 0. It follows that H.c1 c2 / D Hc1 Hc2 D 0 0 D 0 Therefore c1 c2 is also a code word. Property 2 of a parity check code. The code words corresponding to the parity check matrix H D [A B] are identical to the code words corresponding to the parity check matrix Q Q H D [B 1 A; I] D [A I], where I is the r r identity matrix. n r c1 Proof. Let c D be a code word corresponding to the matrix H D [A B], where cn r C1 n cn r are the rst .n r/ components of the vector and cn r C1 are the last r components 1 n of the vector. Then (11.25) Hc D Acn r Bcn r C1 D 0 n 1 Multiplying by B
1
(11.24)
we get B
1
Acn 1
Icn n
r C1
D0
(11.26)
Q or Hc D 0. Q Q From Property 2 we see that parity check matrices of the form H D [A I] are not less general than parity check matrices of the form H D [A B], where det[B] 6D 0. In general, we can consider any r n matrix as a parity check matrix, provided that some set of r columns has a non-zero determinant. If we are not concerned with the order by which the elements of a code word are transmitted, then such a code would be equivalent to a code formed by a parity check matrix of the form H D [A I] (11.27)
The form of the matrix (11.27) is called canonical or systematic form. We assume that the last r columns of H have a non-zero determinant and therefore that the parity check matrix can be expressed in canonical form. Property 3 of a parity check code. There are exactly 2n check code. Proof. Referring to the proof of Property 2, we nd that cn n
r C1 r
D Acn 1
(11.28)
For each of the 2n r D 2k possible binary vectors cn r it is possible to compute the 1 corresponding vector cn r C1 . Each of these code words is unique as all of them differ in n the rst .n r/ D k positions. Assume that there are more than 2k code words; then at least two will agree in the rst .n r/ D k positions. But from (11.28) we nd that these two code words also agree in the last r positions and therefore they are identical.
835
p0 : : : pr
1]
(11.29)
where the rst k D .n r/ bits are called information bits and the last r bits are called parity check bits. As mentioned in Section 11.1, a parity check code that has code words of length n that are obtained by encoding k information bits is an .n; k/ code. Property 4 of a parity check code. A code word of weight w exists if and only if the modulo 2 sum of w columns of H equals 0. Proof. c is a code word if and only if Hc D 0. Let hi be the i-th column of H and let c j be the j-th component of c. Therefore, if c is a code word, then
n X jD1
hjcj D 0
(11.30)
If c is a code word of weight w, then there are exactly w non-zero components of c, for example c j1 ; c j2 ; : : : ; c jw . Consequently h j1 h j2 h jw D 0, thus a code word of weight w implies that the sum of w columns of H equals 0. Conversely, if h j1 h j2 h jw D 0 then Hc D 0, where c is a binary vector with elements equal to 1 in positions j1 ; j2 ; : : : ; jw . From Property 1 of a group code and also from Properties 1 and 4 of a parity check code we obtain the following property.
H Property 5 of a parity check code. A parity check code has minimum distance dmin if H columns of H is equal to 0, but no modulo 2 sum of fewer some modulo 2 sum of dmin H than dmin columns of H is equal to 0. Property 5 may be considered as the fundamental property of parity check codes, as it forms the basis for the design of almost all such codes. An important exception is constituted by low-density parity check codes, which will be discussed in Section 11.7. A limit on the H number of parity check bits required for a given block length n and given dmin derives directly from this property.
Property 6 of a parity check code. A binary parity check code exists of block length n H and minimum distance dmin , having no more than r parity check bits, where 1 0 H dmin 2 j X n 1 k A C1 (11.31) r D log2 @ i i D0 Proof. The proof derives from the following exhaustive construction procedure of the parity check matrix of the code. Step 1: choose as the rst column of H any non-zero vector with r components. Step 2: choose as the second column of H any non-zero vector different from the rst.
836
Step 3: choose as the i-th column of H any vector distinct from all vectors obtained by H modulo 2 sum of .dmin 2/ or fewer previously chosen columns.
H Clearly such a procedure will result in a matrix H where no set of .dmin 1/ or fewer columns of H sum to 0. However, we must show that we can indeed continue this process for n columns. After applying this procedure for .n 1/ columns, there will be at most
Nc .n
H 1; dmin
2/ D 1 C
n 1
n 2
C C
n
H dmin
(11.32)
distinct vectors that are forbidden for the choice of the last column, but there are 2r vectors H to choose from; observing (11.31) and (11.32) we get 2r > Nc .n 1; dmin 2/. Thus n H columns can always be chosen where no set of .dmin 1/ or fewer columns sums to zero. H From Property 5, the code therefore has minimum distance at least dmin .
# D
"
I A
# cn 1
r
D G T cn 1
(11.33)
where G D [I AT ] is a k n binary matrix, and I is the k k identity matrix. Taking the transpose of (11.33), we obtain (11.34) cT D .cn r /T G 1 thus the code words, considered now as row vectors, are given as all linear combinations of the rows of G, which is called the generator matrix of the code. A parity check code can be specied by giving its parity check matrix H or its generator matrix G. Example 11.2.2 Consider the parity check code (7,4) with the parity check matrix 3 1 1 0 1 1 0 0 H D 4 1 1 1 0 0 1 0 5 D [A I] 1 0 1 1 0 0 1 2
(11.35)
Expressing a general code word according to (11.29), to every 4 information bits 3 parity check bits are added, related to the information bits by the equations (see (11.28)) p0 D m 0 m 1 m 3 p1 D m 0 m 1 m 2 p2 D m 0 m 2 m 3 (11.36)
837
(11.37)
There are 16 code words consisting of all linear combinations of the rows of G. By inspection, we nd that the minimum weight of a non-zero code word is 3; hence, from (11.18) H the code has dmin D 3 and therefore is a single error correcting code.
Cosets
The 2n possible binary sequences of length n are partitioned into 2r sets, called cosets, by a group code with 2k D 2n r code words; this partitioning is done as follows: Step 1: choose the rst set as the set of code words c1 ; c2 ; : : : ; c2k . Step 2: choose any vector, say, 2 , that is not a code word; then choose the second set as c 1 2 ; c 2 2 ; : : : ; c 2k 2 . Step i: choose any vector, say, i , not included in any previous set; choose the i-th set, i.e. coset, as c1 i ; c2 i ; : : : ; c2k i . The partitioning continues until all 2n vectors are used. Note that each coset contains 2k vectors; if we show that no vector can appear in more than one coset, we will have demonstrated that there are 2r D 2n k cosets. Property 1 of cosets. Every binary vector of length n appears in one and only one coset.
Proof. Every vector appears in at least one coset as the partitioning stops only when all vectors are used. Suppose that a vector appeared twice in one coset; then for some value of the index i we have c j1 i D c j2 i , or c j1 D c j2 , that is a contradiction as all code words are unique. Suppose that a vector appears in two cosets; then c j1 i1 D c j2 i2 , where we assume i 2 > i 1 . Then i2 D c j1 c j2 i1 D c j3 i1 , that is a contradiction as i2 would have appeared in a previous coset, against the hypothesis.
838
Example 11.2.3 Consider partitioning the 24 binary vectors of length 4 into cosets using the group code with code words 0000, 0011, 1100, 1111, as follows: 0 D 0 D 0 D 1 0 0 1 0 0 0 1 1 0 1 1 0 0 0 0 1 0 0 1 0 1 1 0 0 1 0 0 1 1 1 1 0 1 1 0 1 0 0 1 1 0 1 1 0 1 1 1 0 1 1 0 1 1 1 0 0 1 0 0 1
2 3 4
(11.38)
The vectors 1 D 0; 2 ; 3 ; : : : ; 2r , are called coset leaders; the partitioning (11.38) is called coset table or decoding table. Property 2 of cosets. Suppose that instead of choosing i as the coset leader of the i-th coset, we choose another element of that coset; the new coset formed by using this new coset leader contains exactly the same vectors as the old coset. Proof. Assume that the new coset leader is i c j1 , and that z is an element of the new coset; then z D i c j1 c j2 D i c j3 , so z is an element of the old coset. As the new and the old cosets both contain 2k vectors and all vectors in a coset are unique, every element of the new coset belongs to the old coset and vice versa. Example 11.2.4 Suppose that in the previous example we had chosen the third coset leader as 0100; then the table (11.38) would be 00 2 D 0 0 3 D 0 1 4 D 1 0 0 0 0 1 0 1 0 0 0 0 0 1 0 0 1 0 1 1 1 0 1 0 1 1 1 1 1 0 1 1 0 1 0 0 0 1 0 1 0 0 1 1 1 0 1 1 0 1 1 1 1 0 1 0 1 1
(11.39)
839
Proof. Assume that the received word is the j-th vector in the i-th coset. The received word, given by z D c j i , is corrected to the code word c j and the distance between the received word and the j-th code word is w.i /. Suppose that another code word, say, ck , is closer to the received vector: then w.ck c j i / < w.i / or w.c` i / < w.i / (11.41) but this cannot be as w.i / is assumed to be the minimum weight vector in its coset and c` i is in that coset. We note that the coset leaders determine the only error patterns that can be corrected by the code. Coset leaders, moreover, have many other interesting properties: for example, H if a code has minimum distance dmin , all binary n-tuple of weight less than or equal to j H k dmin 1 are coset leaders. 2 Denition 11.5 A code for which coset leaders are all vectors of weight t or less, and no others, is called a perfect t-error correcting code. A code for which coset leaders are all vectors of weight t or less, and some vectors of weight t C 1 but not all, and no others, is called quasi-perfect t-error correcting code. The perfect binary codes are: 1. codes given by the repetition of n bits, with n odd: these codes contain only two code words, 000 : : : 0 (all zeros) and 111 : : : 1 (all ones), and correct t D .n 1/=2 H errors .dmin D n/;
H 2. Hamming codes: these codes correct t D 1 errors .dmin D 3/ and have n D 2r 1, k D n r, r > 1; the columns of the matrix H are given by all non-zero vectors of length r; H 3. Golay code: t D 3 .dmin D 7/, n D 23, k D 12, r D 11.
(11.40)
The following modication of the decoding method dealt with in this section will be useful later on: Step 10 : locate the received vector in the coset table and identify the coset leader of the coset containing that vector. Step 20 : add the coset leader to the received vector to nd the decoded code word.
Syndrome decoding
A third method of decoding is based on the concept of syndrome. Among the methods described in this section, syndrome decoding is the only method of practical value for a code with a large number of code words.
840
Denition 11.6 For any parity check matrix H, we dene the syndrome s.z/ of a binary vector z of length n as s.z/ D Hz (11.42) We note that the syndrome is a vector of length r, whereas z is a vector of length n. Therefore many vectors will have the same syndrome. All code words have an all zero syndrome and these are the only vectors with this property. This property of the code words is a special case of the following: Property 3 of cosets. All vectors in the same coset have the same syndrome; vectors in different cosets have distinct syndromes. Proof. Assume that z1 and z2 are in the same coset, say, the i-th: then z1 D i c j1 and z2 D i c j2 . Moreover s.z1 / D Hz1 D H.i c j1 / D Hi Hc j1 D Hi 0 D s.i /. Similarly s.z2 / D s.i /, so s.z1 / D s.z2 / D s.i /: this proves the rst part of the property. Now assume that z1 and z2 are in different cosets, say, the i 1 -th and i 2 -th: then z1 D i1 c j1 and z2 D i2 c j2 , so s.z1 / D s.i1 / and s.z2 / D s.i2 /. If s.z1 / D s.z2 / then s.i1 / D s.i2 /, which implies Hi1 D Hi2 . Consequently H.i1 i2 / D 0, or i1 i2 is a code word, say c j3 . Then i2 D i1 c j3 , which implies that i1 and i2 are in the same coset, that is a contradiction. Thus the assumption that s.z1 / D s.z2 / is incorrect. From Property 3 we see that there is a one-to-one relation between cosets and syndromes; this leads to the third method of decoding, which proceeds as follows: Step 100 : compute the syndrome of the received vector; this syndrome identies the coset in which the received vector is. Identify then the leader of that coset. Step 200 : add the coset leader to the received vector to nd the decoded code word. Example 11.2.5 Consider the parity check matrix 3 1 1 0 1 0 0 HD41 0 1 0 1 05 1 1 1 0 0 1 2
(11.43)
The coset leaders and their respective syndromes obtained using (11.42) are reported in Table 11.1. Suppose that the vector z D 000111 is received. To decode we rst compute the syndrome 2 3 1 607 2 3 6 7 1 607 Hz D 4 1 5, then by Table 11.1 we identify the coset leader as 6 7, and obtain the decoded 607 6 7 1 405 0 code word . 1 0 0 1 1 1 /.
841
Table 11.1 Coset leaders and respective syndromes for Example 11.2.5.
Coset leader 000000 000001 000010 000100 001000 010000 100000 100001
The advantage of syndrome decoding over the other decoding methods previously described is that there is no need to memorize the entire decoding table at the receiver. The rst part of Step 100 , namely computing the syndrome, is trivial. The second part of Step 100 , namely identifying the coset leader corresponding to that syndrome, is the difcult part of the procedure; in general it requires a RAM with 2r memory locations, addressed by the syndrome of r bits and containing the coset leaders of n bits. Overall the memory bits are n2r . There is also an algebraic method to identify the coset leader. In fact, this problem is equivalent to nding the minimum set of columns of the parity check matrix which sum to the syndrome. In other words, we must nd the vector z of minimum weight such that Hz D s. For a single error correcting Hamming code, all coset leaders are of weight 1 or 0, so a non-zero syndrome corresponds to a single column of H and the correspondence between syndrome and coset leader is simple. For a code with coset leaders of weight 0, 1, or 2, the syndrome is either 0, a single column of H, or the sum of two columns, etc. For a particular class of codes that will be considered later, the structure of the construction of H will allow identication of the coset leader starting from the syndrome by using algebraic procedures. In general, each class of codes leads to a different technique to perform this task. Property 7 of parity check codes. There are exactly 2r correctable error vectors for a parity check code with r parity check bits. Proof. Correctable error vectors are given by the coset leaders and there are 2n k D 2r of them, all of which are distinct. On the other hand, there are 2r distinct syndromes and each corresponds to a correctable error vector. For a binary symmetric channel (see Denition 6.1) we should correct all error vectors of weight i, i D 0; 1; 2; : : : , until we exhaust the capability of the code. Specically, we should try to use a perfect code or a quasi-perfect code. For a quasi-perfect t-error correcting code, the coset leaders consist of all error vectors of weight i D 0; 1; 2; : : : ; t, and some vectors of weight t C 1. Nonbinary parity check codes are discussed in Appendix 11.A.
842
11.2.2
Fundamentals of algebra
The calculation of parity check bits from information bits involves solving linear equations. This procedure is particularly easy for binary codes since we use modulo-2 arithmetic. An obvious question is whether or not the concepts of the previous section generalize to codes with symbols taken from alphabets with a larger cardinality, say, alphabets with q symbols. We will see that the answer can be yes or no according to the value of q; furthermore, even if the answer is yes we might not be able to use modulo q arithmetic. Consider the equation for the unknown x ax D b (11.44)
where a and b are known coefcients, and all values are from the nite alphabet f0; 1; 2, : : : ; q 1g. First, we need to introduce the concept of multiplication, which is normally given in the form of a multiplication table, as the one given in Table 11.2 for the three elements f0; 1; 2g. Table 11.2 allows us to solve (11.44) for any values of a and b, except a D 0. For example, the solution to equation 2x D 1 is x D 2, as from the multiplication table we nd 2 2 D 1. Let us now consider the case of an alphabet with four elements. A multiplication table for the four elements f0; 1; 2; 3g, resulting from the modulo 4 arithmetic, is given in Table 11.3. Note that the equation 2x D 2 has two solutions, x D 1 and x D 3, and equation 2x D 1 has no solution. It is possible to construct a multiplication table that allows the equation (11.44) to be solved uniquely for x, provided that a 6D 0, as shown in Table 11.4.
Table 11.2 Multiplication table for 3 elements (modulo 3 arithmetic).
0 1 2
0 0 0 0
1 0 1 2
2 0 2 1
Table 11.3 Multiplication table for an alphabet with four elements (modulo 4 arithmetic).
0 1 2 3
0 0 0 0 0
1 0 1 2 3
2 0 2 0 2
3 0 3 2 1
843
0 1 2 3
0 0 0 0 0
1 0 1 2 3
2 0 2 3 1
3 0 3 1 2
Note that Table 11.4 is not obtained using modulo 4 arithmetic. For example, 2x D 3 has the solution x D 2, and 2x D 1 has the solution x D 3.
Modulo q arithmetic
Consider the elements f0; 1; 2; : : : ; q 1g, where q is a positive integer larger than or equal to 2. We dene two operations for combining pairs of elements from this set. The rst, denoted by , is called modulo q addition and is dened as ( aCb if 0 a C b < q c Dab D (11.45) aCb q if a C b q Here a C b is the ordinary addition operation for integers that may produce an integer not in the set. In this case q is subtracted from a C b and a C b q is always an element in the set f0; 1; 2; : : : ; q 1g. The second operation, denoted by , is called modulo q multiplication and is dened as 8 > ab if 0 ab < q < j ab k (11.46) d Da bD > ab q if ab q : q j k ab Note that ab q q, is the remainder or residue of the division of ab by q, and is always an integer in the set f0; 1; 2; : : : ; q 1g. Often we will omit the notation and write a b simply as ab. We recall that special names are given to sets which possess certain properties with respect to operations. Consider the general set G that contains the elements f; ; ; ; : : : g, and two operations for combining elements from the set. We denote the rst operation 4 (addition), and the second operation (multiplication). Often we will omit the notation and write ab simply as ab. The properties we are interested in are: 1. Existence of additive identity. For every 2 G, there exists an element ; 2 G, called additive identity, such that 4; D ;4 D . 2. Existence of additive inverse. For every 2 G, there exists an element 2 G, called additive inverse of , and indicated by , such that 4 D 4 D ;.
844
3. Additive closure. For every ; 2 G, not necessarily distinct, 4 2 G. 4. Additive associative law. For every ; ; 2 G, 4.4 / D .4/4 .
5. Additive commutative law. For every ; 2 G, 4 D 4. 6. Multiplicative closure. For every ; 2 G, not necessarily distinct, 2 G. 7. Multiplicative associative law. For every ; ; 8. Distributive law. For every ; ; . /4. /. 2 G, . / D ./ .
9. Multiplicative commutative law. For every ; 2 G, D . 10. Existence of multiplicative identity. For every 2 G, there exists an element I 2 G, called multiplicative identity, such that I D I D . 11. Existence of multiplicative inverse. For every 2 G, except the element ;, there exists an element 2 G, called multiplicative inverse of , and indicated with 1 , such that D D I . Any set G for which Properties 1 4 hold is called a group with respect to 4. If G has a nite number of elements, then G is called nite group and the number of elements of G is called the order of G. Any set G for which Properties 1 5 hold is called an Abelian group with respect to 4. Any set G for which Properties 1 8 hold is called a ring with respect to the operations 4 and . Any set G for which Properties 1 9 hold is called a commutative ring with respect to the operations 4 and . Any set G for which Properties 1 10 hold is called a commutative ring with identity. Any set G for which Properties 1 11 hold is called a eld. It can be seen that the set f0; 1; 2; : : : ; q 1g is a commutative ring with identity with respect to the operations of addition dened in (11.45) and multiplication dened in (11.46). We will show by the next three properties that this set satises also Property 11 if and only if q is a prime: in other words, we will show that the set f0; 1; 2; : : : ; q 1g is a eld with respect to the modulo q addition and modulo q multiplication if and only if q is a prime. Finite elds are called Galois elds; a eld of q elements is usually denoted as G F.q/. Property 11a of modulo q arithmetic. If q is not a prime, each factor of q (less than q and greater than 1) does not have a multiplicative inverse. Proof. Let q D ab, where 1 < a; b < q; then, observing (11.46), a b D 0. Assume that a has a multiplicative inverse a 1 ; then a 1 .a b/ D a 1 0 D 0. Now, from a 1 .a b/ D 0 it is 1 b D 0; this implies b D 0, which is a contradiction as b > 1. Similarly we show that b does not have a multiplicative inverse.
845
b D 0, then a D 0, or b D 0,
Proof. Assume a b D 0 and a; b > 0; then ab D K q, where K < min.a; b/. If 1 < a q 1 and a has no factors in common with q, then it must divide K ; but this is impossible as K < min.a; b/. The only other possibility is that a D 1, but then a b 6D 0 as ab < q. Property 11c of modulo q arithmetic. If q is a prime, all non-zero elements of the set f0; 1; 2; : : : ; q 1g have multiplicative inverse. Proof. Assume the converse, that is the element j, with 1 j q 1, does not have a multiplicative inverse; then there must be two distinct elements a; b 2 f0; 1; 2; : : : ; q 1g such that a j D b j. This is a consequence of the fact that the product i j can only assume values in the set f0; 2; 3; : : : ; q 1g, as by assumption i j 6D 1; then .a On the other hand, q .b j/ .q b/ .a .q .b j, and b// j D0 (11.48) b/ D 0. This implies j// D 0 (11.47) j/ D .q
But j 6D 0 and consequently, by Property 11b, we have a .q a D b, which is a contradiction. Denition 11.7 An ideal I is a subset of elements of a ring R such that:
1. I is a subgroup of the additive group R, that is the elements of I form a group with respect to the addition dened in R; 2. for any element a of I and any element r of R, ar and ra are in I .
846
We assume that gm 6D 0 and f n 6D 0. We dene m as degree of the polynomial g.y/, and we write m D deg.g.y//; in particular, if g.y/ D a, a 2 f0; 1; 2; : : : ; p 1g, we say that deg.g.y// D 0. Similarly, it is n D deg. f .y//. If gm D 1, we say that g.y/ is a monic polynomial. Assume m n: then the addition among polynomials is dened as f .y/ C g.y/ D . f 0 g0 / C . f 1 g1 /y C . f 2 g2 /y 2 C C . f m gm / y m C C f n y n (11.50) Example 11.2.6 Let p D 5, f .y/ D 1C3y C2y 4 , and g.y/ D 4C3y C3y 2 ; then f .y/Cg.y/ D y C3y 2 C2y 4 . Note that deg. f .y/ C g.y// max.deg. f .y//; deg.g.y/// Multiplication among polynomials is dened as usual f .y/ g.y/ D d0 C d1 y C C dmCn y mCn (11.51)
where the arithmetic to perform operations with the various coefcients is modulo p, di D . f 0 gi / . f 1 gi
1 / . fi 1
g1 / . f i
g0 /
(11.52)
Example 11.2.7 Let p D 2, f .y/ D 1 C y C y 3 , and g.y/ D 1 C y 2 C y 3 ; then f .y/ g.y/ D 1 C y C y 2 C y3 C y4 C y5 C y6. Note that deg. f .y/ g.y// D deg. f .y// C deg.g.y// Denition 11.8 If f .y/ g.y/ D d.y/, we say that f .y/ divides d.y/, and g.y/ divides d.y/. We say that p.y/ is an irreducible polynomial if and only if, assuming another polynomial a.y/ divides p.y/, then a.y/ D a 2 f0; 1; : : : ; p 1g or a.y/ D k p.y/, with k 2 f0; 1; : : : ; p 1g. The concept of an irreducible polynomial plays the same role in the theory of polynomials as does the concept of a prime number in the number theory.
847
Example 11.2.8 Let p D 2 and q.y/ D 1 C y C y 3 ; then the set P consists of 23 polynomials, f0; 1; y; y C 1; y 2 ; y 2 C 1; y 2 C y; y 2 C y C 1g. Example 11.2.9 Let p D 3 and q.y/ D 2y 2 ; then the set P consists of 32 polynomials, f0; 1; 2; y; y C 1; y+2; 2y; 2y C 1; 2y C 2g. We now dene two operations among polynomials of the set P, namely modulo q.y/ addition, denoted by 4, and modulo q.y/ multiplication, denoted by . Modulo q.y/ addition is dened for every pair of polynomials a.y/ and b.y/ from the set P as a.y/4b.y/ D a.y/ C b.y/ (11.53)
where a.y/ C b.y/ is dened in (11.50). The denition of modulo q.y/ multiplication requires the knowledge of the Euclidean division algorithm. Euclidean division algorithm. For every pair of polynomials .y/ and .y/ with coefcients from some eld, and deg..y// deg..y// > 0, there exists a unique pair of polynomials q.y/ and r.y/ such that .y/ D q.y/ .y/ C r.y/ (11.54)
where 0 deg.r.y// < deg..y//; polynomials q.y/ and r.y/ are called, respectively, quotient polynomial and remainder or residue polynomial. In a notation analogous to that used for integers we can write j .y/ k (11.55) q.y/ D .y/ and r.y/ D .y/ j .y/ k .y/ .y/ (11.56)
Example 11.2.10 Let p D 2, .y/ D y 4 C 1, and .y/ D y 3 C y C 1; then y 4 C 1 D y.y 3 C y C 1/ C y 2 C y C 1, so q.y/ D y and r.y/ D y 2 C y C 1. We dene modulo q.y/ multiplication, denoted by , for polynomials a.y/ and b.y/ in the set P as 8 > a.y/ b.y/ if deg.a.y/ b.y// < deg.q.y// < j a.y/ b.y/ k a.y/b.y/ D > a.y/ b.y/ q.y/ otherwise : q.y/ (11.57)
848
It is easier to think of (11.57) as a typical multiplication operation for polynomials whose coefcients are given according to (11.52). If in this multiplication there are terms of degree greater than or equal to deg.q.y//, then we use the relation q.y/ D 0 to lower the degree. Example 11.2.11 Let p D 2 and q.y/ D 1 C y C y 3 ; then .y 2 C 1/.y C 1/ D y 3 C y 2 C y C 1 D . 1 y/ C y 2 C y C 1 D .1 C y/ C y 2 C y C 1 D y 2 . It can be shown that the set of polynomials with coefcients from some eld and degree less than deg.q.y// is a commutative ring with identity with respect to the operations modulo q.y/ addition and modulo q.y/ multiplication. We now nd under what conditions this set of polynomials and operations forms a eld. Property 11a of modular polynomial arithmetic. If q.y/ is not irreducible, then the factors of q.y/, of degree greater than zero and less than deg.q.y//, do not have multiplicative inverses. Proof. Let q.y/ D a.y/ b.y/, where 0 < deg.a.y//; deg.b.y// < deg.q.y//; then a.y/b.y/ D 0. Assume a.y/ has a multiplicative inverse, a 1 .y/; then, from a 1 .y/ .a.y/b.y// D a 1 .y/0 D 0 it is .a 1 .y/a.y//b.y/ D 0, then 1b.y/ D 0, or b.y/ D 0. The last equation is a contradiction as by assumption deg.b.y// > 0. Similarly, we show that b.y/ does not have a multiplicative inverse. We give without proof the following properties. Property 11b of modular polynomial arithmetic. If q.y/ is irreducible and a.y/b.y/ D 0, then a.y/ D 0, or b.y/ D 0, or a.y/ D b.y/ D 0. Property 11c of modular polynomial arithmetic. If q.y/ is irreducible, all non-zero elements of the set of polynomials P of degree less than deg.q.y// have multiplicative inverses. We now have that the set of polynomials with coefcients from some eld and degree less than deg.q.y// forms a eld, with respect to the operations of modulo q.y/ addition and modulo q.y/ multiplication, if and only if q.y/ is irreducible. Furthermore it can be shown that there exists at least one irreducible polynomial of degree m, for every m 1, with coefcients from a generic eld f0; 1; 2; : : : ; p 1g. We now have a method of generating a eld with pm elements. Example 11.2.12 Let p D 2 and q.y/ D y 2 C y C 1; we have that q.y/ is irreducible. Consider the set P with elements f0; 1; y; y C 1g. The addition and multiplication tables for these elements modulo y 2 C y C 1 are given in Table 11.5 and Table 11.6, respectively.
849
4 0 1 y yC1
0 0 1 y yC1
1 1 0 yC1 y
y y yC1 0 1
yC1 yC1 y 1 0
0 1 y yC1
0 0 0 0 0
1 0 1 y yC1
y 0 y yC1 1
yC1 0 yC1 1 y
qi y i
qi 2 G F. p/
(11.58)
ai y i
ai 2 G F. p/
(11.59)
bi y i
bi 2 G F. p/
(11.60)
(11.61)
is illustrated in Figure 11.1. The implementation of a device to perform the multiplication is slightly more complicated, as illustrated in Figure 11.2, where Tc is the period of the clock applied to the shift-register (SR) with m elements, and all operations are modulo p. Let us dene d.y/ D
m 1 X i D0
(11.62)
850
a0
b0 mod p s0
am1
bm1 mod p
s m1
Figure 11.1. Device for the sum of two elements .a0 ; : : : ; am 1 / and .b0 ; : : : ; bm 1 / of GF.pm /.
Figure 11.2. Device for the multiplication of two elements .a0 ; : : : ; am 1 / and .b0 ; : : : ; bm 1 / of GF.pm /. Tc is the clock period, and ACC denotes an accumulator. All additions and multiplications are modulo p.
The device is based on the following decomposition a.y/ b.y/ mod q.y/ D
m 1 X i D0
(11.63)
851
where additions and multiplications are modulo p. Now, using the identity 0 mod q.y/, note that the following relation holds: y b.y/ D b0 y C b1 y 2 C C bm D . bm
1 qm 1 2y m 1
i D0
qi y i D
C bm
1
1y
ym 1 : (11.64) The term .y i b.y// mod q.y/ is thus obtained by initializing the SR of Figure 11.2 to the sequence .b0 ; : : : ; bm 1 /, and by applying i clock pulses; the desired result is then contained in the shift register. Observing (11.63), we nd that it is necessary to multiply each element of the SR by ai and accumulate the result; after multiplications by all coefcients fai g have been performed, the nal result is given by the content of the accumulators. Note that in the binary case, for p D 2, the operations of addition and multiplication are carried out by XOR and AND functions, respectively. bm
1 qm 2
q0 / C .b0
q1 / y C C .bm
bm
1 qm
qm
1/
We observe that j
i times
i times
j times
(11.66)
852
Substituting (11.65) in (11.66), and observing that has a multiplicative inverse, we can multiply from the right by this inverse to obtain j j i D i times D 1 (11.67)
Denition 11.9 For every non-zero eld element, , the order of is the smallest integer ` such that ` D 1. Example 11.2.13 Consider the eld with elements f0; 1; 2; 3; 4g, and modulo 5 arithmetic. Then element 1 2 3 4 order 1 4 4 2
(11.68)
Example 11.2.14 Consider the eld G F.22 / with 4 elements, f0; 1; y; y C 1g, and addition and multiplication modulo y 2 C y C 1. Then element 1 y yC1 order 1 3 3
(11.69)
6. An element from the eld G F.q/ is said to be primitive if it has order q 1. For elds generated by arithmetic modulo a polynomial q.y/, if the eld element y is primitive we say that q.y/ is a primitive irreducible polynomial. A property of nite elds that we give without proof is that every nite eld has at least one primitive element; we note that once a primitive element has been identied, every other non-zero eld element can be obtained by multiplying the primitive element by itself an appropriate number of times. A list of primitive polynomials for the ground eld G F.2/ is given in Table 11.7. Example 11.2.15 For the eld G F.4/ generated by the polynomial arithmetic modulo q.y/ D y 2 C y C 1, for the ground eld G F.2/, y is a primitive element (see (11.69)); thus y 2 C y C 1 is a primitive polynomial. 7. The order of every non-zero element of G F.q/ must divide .q 1/.
853
Table 11.7 List of primitive polynomials q.y/ of degree m for the ground eld GF.2/.
m 2 3 4 5 6 7 8 9 10 11 12 13 1 C y C y2 1 C y C y3 1 C y C y4 1 C y2 C y5 1 C y C y6 1 C y3 C y7 1 C y2 C y3 C y4 C y8 1 C y4 C y9 1 C y 3 C y 10 1 C y 2 C y 11 1 C y C y 4 C y 6 C y 12 1 C y C y 3 C y 4 C y 13
m 14 15 16 17 18 19 20 21 22 23 24 1 C y C y 6 C y 10 C y 14 1 C y C y 15 1 C y C y 3 C y 12 C y 16 1 C y 3 C y 17 1 C y 7 C y 18 1 C y C y 2 C y 5 C y 19 1 C y 3 C y 20 1 C y 2 C y 21 1 C y C y 22 1 C y 5 C y 23 1 C y C y 2 C y 7 C y 24
Proof. Every non-zero element can be written as the power of a primitive element p ; this implies that there is some i < .q 1/ such that D p i times p p D ip
q 1 j
(11.70)
Note that from the denition of a primitive element we get p D 1, but p 6D 1 for j < .q 1/; furthermore there exists an integer ` such that ` D ip` D 1. Consequently .`/.i/ is a multiple of .q 1/ and it is exactly the smallest multiple of i that is a multiple of .q 1/, thus .i/.`/ D l:c:m:.i; q 1/, i.e. the least common multiple of i and .q 1/. We recall that ab (11.71) l:c:m:.a; b/ D g:c:d:.a; b/ where g:c:d:.a; b/ is the greatest common divisor of a and b. Thus .i/.`/ D and .i/.q 1/ g:c:d:.i; q 1/ D` (11.72)
q 1 g:c:d:.i; q
1/
(11.73)
Example 11.2.16 Let p be a primitive element of G F.16/; from (11.73) the orders of the non-zero eld elements are:
854
eld element D i p 2 p 3 p 4 p 5 p 6 p 7 p 8 p 9 p 10 p 11 p 12 p 13 p 14 p 15 p
g:c:d:.i; q
1/
1 1 3 1 5 3 1 1 3 5 1 3 1 1 15
8. A ground eld can itself be generated as an extension eld. For example G F.16/ can be generated by taking an irreducible polynomial of degree 4 with coefcients from G F.2/, which we would call G F.24 /, or by taking an irreducible polynomial of degree 2 with coefcients from G F.4/, which we would call G F.42 /. In either case we would have the same eld, except for the names of the elements. Example 11.2.17 Consider the eld G F.23 / generated by the primitive polynomial q.y/ D 1 C y C y 3 , with ground eld G F.2/. As q.y/ is a primitive polynomial, each element of G F.23 /, except the zero element, can be expressed as a power of y. Recalling the polynomial representation P, we may attach to each polynomial a vector representation, with m components on G F. p/ given by the coefcients of the powers of the variable y. The three representations are reported in Table 11.8.
Roots of a polynomial
Consider a polynomial of degree m with coefcients that are elements of some eld. We will use the variable x, as the polynomials are now considered for a purpose that is not that of generating a nite eld. In fact, the eld of the coefcients may itself have a polynomial representation.
855
Exponential
Polynomial
Binary (y 0 y 1 y 2 ) 0 1 0 0 1 0 1 1 0 0 1 0 1 1 1 0 0 0 0 1 0 1 1 1
0 1 y y2 y3 y4 y5 y6
0 1 y y2 1Cy y C y2 1 C y C y2 1 C y2
Consider, for example, a polynomial in x with coefcients from G F.4/. We immediately see that it is not worth using the notation f0; 1; y; yC1g to identify the 4 elements of G F.4/, as the notation f0; 1; ; g would be much simpler. For example, a polynomial of degree three with coefcients from G F.4/ is given by f .x/ D x 3 C x 2 C 1. Given any polynomial f .x/, we say that is a root of the equation f .x/ D 0 or, more simply, that it is a root of f .x/, if and only if f . / D 0. The denition is more complicated than it appears, as we must know the meaning of the two members of the equation f . / D 0. For example, we recall that the fundamental theorem of algebra states that every polynomial of degree m has exactly m roots, not necessarily distinct. If we take the polynomial f .x/ D x 2 C x C 1 with coefcients from f0; 1g, what are its roots? As f .0/ D f .1/ D 1, we have that neither 0 nor 1 are roots. Before proceeding, we recall a similar situation that we encounter in ordinary algebra. The polynomial x 2 C 3, with coefcients in the eld of real numbers, has two roots in the eld of complex numbers; however, no roots exist in the eld of real numbers; therefore the polynomial does not have factors whose coefcients are real numbers. Thus we would say that the polynomial is irreducible, yet even the irreducible polynomial has complex-valued roots and can be factorized. This situation is due to the fact that, if we have a polynomial f .x/ with coefcients from some eld, the roots of the polynomial are either from that eld or from an extension eld of that eld. For example, take the polynomial f .x/ D x 2 C x C 1 with coefcients from G F.2/, and consider the extension eld G F.4/ with elements f0; 1; ; g that obey the addition and the multiplication rules given in Table 11.9 and Table 11.10, respectively. Then f ./ D f .x/jxD D 2 441 D .4/41 D 141 D 0, thus is a root. Similarly we nd f ./ D f .x/jxD D 2 441 D .4/41 D 141 D 0, thus the two roots of f .x/ are and . We can factor f .x/ into two factors, each of which is a polynomial in x with coefcients from G F.4/. For this purpose we consider .x4 /.x4 / D .x4/.x4/; leaving out the notations 4 and for C and we get .x4/.x4/ D x 2 C . C / x C D x 2 C x C 1 (11.75)
856
4 0 1
0 0 1
1 1 0
0 1
1 0
0 1
0 0 0 0 0
1 0 1
0 1
0 1
Thus if we use the operations dened in G F.4/, .x C/ and .x C/ are factors of x 2 Cx C1; it remains that x 2 C x C 1 is irreducible as it has no factors with coefcients from G F.2/. Property 1 of the roots of a polynomial. .x C . //, is a factor of f .x/. If is a root of f .x/ D 0, then .x / to get (11.76) /, that is
Proof. Using the Euclidean division algorithm, we divide f .x/ by .x f .x/ D Q.x/ .x where deg.r.x// < deg.x / D 1. Therefore f .x/ D Q.x/ .x But f . / D 0, so therefore f .x/ D Q.x/ .x / f . / D 0 D Q. / . / C r0 / C r0 D r0 / C r.x/
(11.77)
(11.78)
(11.79)
Property 2 of the roots of a polynomial. If f .x/ is an arbitrary polynomial with coefcients from G F. p/, p a prime, and is a root of f .x/, then p is also a root of f .x/. Proof. We consider the polynomial f .x/ D f 0 C f 1 xC f 2 x 2 C C f m x m , where f i 2 G F. p/, and form the power . f .x// p . It results . f .x// p D . f 0 C f 1 x C f 2 x 2 C C f m x m / p D f 0 C f 1 x p C C f m x mp
p p p
(11.80)
857
as the cross-terms contain a factor p, which is the same as 0 in G F. p/. On the other hand, p for f i 6D 0, f i D f i , as from Property 7 on page 853 the order of any non-zero element divides p 1; the equation is true also if f i is the zero element. Therefore . f .x// p D f .x p / (11.81)
If is a root of f .x/ D 0, then f ./ D 0, and f p ./ D 0. But f p ./ D f . p /, so that f . p / D 0; therefore p is also a root of f .x/. A more general form of the property just introduced, that we will give without proof, is expressed by the following property. Property 2a of the roots of a polynomial. If f .x/ is an arbitrary polynomial having coefcients from G F.q/, with q a prime or a power of a prime, and is a root of f .x/ D 0, then q is also a root of f .x/ D 0, Example 11.2.18 Consider the polynomial x 2 C x C 1 with coefcients from G F.2/. We already have seen that , element of G F.4/, is a root of x 2 C x C 1 D 0. Therefore 2 is also a root; but 2 D , so is a second root. The polynomial has degree two, thus it has two roots and they are and , as previously seen. Note also that 2 is also a root, but 2 D .
Minimum function
Denition 11.10 Let be an element of an extension eld of G F.q/; the minimum function of , m .x/, is the monic polynomial of least degree with coefcients from G F.q/ such that m .x/jxD D 0. We now list some properties of the minimum function. 1. The minimum function is unique. Proof. Assume there were two minimum functions, of the same degree and monic, m .x/ and m 0 .x/. Form the new polynomial .m .x/ m 0 .x// whose degree is less than the degree of m .x/ and m 0 .x/; but .m .x/ m 0 .x//jxD D 0, so we have a new polynomial, whose degree is less than that of the minimum function, that admits as root. Multiplying by a constant we can thus nd a monic polynomial with this property, but this cannot be since the minimum function is the monic polynomial of least degree for which is a root. 2. The minimum function is irreducible. Proof. Assume the converse were true, that is m .x/ D a.x/ b.x/; then m .x/jxD D a./ b./ D 0. Then either a./ D 0 or b./ D 0, so that is a root of a polynomial of degree less than the degree of m .x/. By making this polynomial monic we arrive at a contradiction. 3. Let f .x/ be any polynomial with coefcients from G F.q/, and let f .x/jxD D 0; then f .x/ is divisible by m .x/.
858
Proof. Use the Euclidean division algorithm to yield f .x/ D Q.x/ m .x/ C r.x/ where deg.r.x// < deg.m .x//. Then we have that f ./ D Q./ m ./ C r./ (11.83) (11.82)
but as f ./ D 0 and m ./ D 0, then r./ D 0. As deg.r.x// < deg.m .x//, the only possibility is r.x/ D 0; thus f .x/ D Q.x/ m .x/. 4. Let f .x/ be any irreducible monic polynomial with coefcients from G F.q/ for which f ./ D 0, where is an element of some extension eld of G F.q/; then f .x/ D m .x/. Proof. From Property 3 f .x/ must be divisible by m .x/, but f .x/ is irreducible, so it is only trivially divisible by m .x/, that is f .x/ D K m .x/: but f .x/ and m .x/ are both monic polynomials, therefore K D 1. We now introduce some interesting propositions. 1. Let be an element of G F.q m /, with q prime; then the polynomial F.x/, dened as F.x/ D
m 1 Y i D0
.x
q / D .x
/ .x
q / .x
q / .x
m 1
(11.84)
has all its coefcients from G F.q/. Proof. Observing Property 7 on page 853, we have that the order of divides q m m therefore q D . Thus we can express F.x/ as F.x/ D therefore F.x q / D
m Y i D1 m Y i D1
1,
.x
q /
(11.85)
.x q
q / D
m Y i D1
.x
q Pm
i 1
/q D
m 1 Y jD0
.x
q /q D .F.x//q
(11.86)
i D0
f i x i ; then
q
m X i D0
fi x i
(11.87)
m X i D0
!q fi x i D
m X i D0
fi x i
(11.88)
859
Equating like coefcients in (11.87) and (11.88) we get f i D f i ; hence f i is a root of the equation x q x D 0. But on the basis of Property 7 on page 853 the q elements from G F.q/ all satisfy the equation x q x D 0, and this equation only has q roots; therefore the coefcients f i are elements from G F.q/. 2. If g.x/ is an irreducible polynomial of degree m with coefcients from G F.q/, and 2 g./ D 0, where is an element of some extension eld of G F.q/, then ; q ; q , : : : , m 1 q are all the roots of g.x/. Proof. At least one root of g.x/ is in G F.q m /; this follows by observing that, if we form G F.q m / using the arithmetic modulo g.y/, then y will be a root of g.x/ D 0. From Qm i Proposition 1, if is an element from G F.q m / then F.x/ D i D01 .x q / has all coefcients from G F.q/; thus F.x/ has degree m, and F./ D 0. As g.x/ is irreducible, we know that g.x/ D K m .x/; but as F./ D 0, and F.x/ and g.x/ have the same degree, 2 m 1 then F.x/ D K 1 m .x/, and therefore g.x/ D K 2 F.x/. As ; q ; q ; : : : ; q , are all roots of F.x/, then they must also be all the roots of g.x/. 3. Let g.x/ be a polynomial with coefcients from G F.q/ which is also irreducible in this eld. Moreover, let g./ D 0, where is an element of some extension eld of G F.q/; then the degree of g.x/ equals the smallest integer k such that q D
2 k 1 k
(11.89)
Proof. We have that deg.g.x// k as ; q ; q ; : : : ; q , are all roots of g.x/ and by assumption are distinct. Assume that deg.g.x// > k; from Proposition 2, we know that must be at least a double root of g.x/ D 0, and therefore g 0 .x/ D .d=dx/g.x/ D 0 must also have as a root. As g.x/ is irreducible we have that g.x/ D K m .x/, but m .x/ must divide g 0 .x/; we get a contradiction because deg.g 0 .x// < deg.g.x//.
860
Example 11.2.20 Consider the eld G F.23 / of the previous example; as .y C 1/, .y C 1/2 D y 2 C 1, .y C 1/4 D y 2 C y C 1, .y C 1/8 D y C 1, the minimum function has degree three; as the minimum function is monic and irreducible, we have m yC1 .x/ D m 3 x 3 C m 2 x 2 C m 1 x C m 0 D x 3 C m 2 x 2 C m 1 x C 1 As m yC1 .y C 1/ D 0, then .y C 1/3 C m 2 .y C 1/2 C m 1 .y C 1/ C 1 D 0 that can be written as y 2 .1 C m 2 / C ym 1 C .m 2 C m 1 C 1/ D 0 (11.93) (11.92) (11.91)
As all coefcients of the powers of y must be zero, we get a system of equations in the unknown m 1 and m 2 , whose solution is given by m 1 D 0 and m 2 D 1. Substitution of this solution in (11.91) yields m yC1 .x/ D x 3 C x 2 C 1 3. Using the minimum function of the multiplicative inverse. Denition 11.11 The reciprocal polynomial of any polynomial m .x/ D m 0 C m 1 x C m 2 x 2 C C m K x K is dened by m .x/ D m 0 x K C m 1 x K 1 C C m K 1 x C m K . We use the following proposition that we give without proof. The minimum function of the multiplicative inverse of a given element is equal to the reciprocal of the minimum function of the given element. In formulae: let D 1, then m .x/ D m .x/. Example 11.2.21 Consider the eld G F.26 / obtained by taking the polynomial arithmetic modulo the irreducible polynomial y 6 C y C 1 with coefcients from G F.2/; the polynomial y 6 C y C 1 is primitive, thus from Property 7 on page 853 any non-zero eld element can be written as a power of the primitive element y. From Proposition 2, we have that the minimum function of y is also the minimum function of y 2 ; y 4 ; y 8 ; y 16 ; y 32 , the minimum function of y 3 is also the minimum function of y 6 ; y 12 ; y 24 ; y 48 ; y 33 , and so forth. We list in Table 11.11 the powers of y that have the same minimum function. Given the minimum function of y 11 , m y 11 D x 6 C x 5 C x 3 C x 2 C 1, we want to nd the minimum function of y 13 . From Table 11.11 we note that y 13 has the same minimum function as y 52 ; furthermore we note that y 52 is the multiplicative inverse of y 11 , as .y 11 /.y 52 / D y 63 D 1. Therefore the minimum function of y 13 is the reciprocal polynomial of m y 11 , given by m y 13 D x 6 C x 4 C x 3 C x C 1. (11.94)
861
Table 11.11 Powers of a primitive element in GF.26 / with the same minimum function.
1 3 5 7 9 11 13 15 21 23 27 31
2 6 10 14 18 22 26 30 42 46 54 62
4 12 20 28 36 44 52 60 29 45 61
8 24 40 56 25 41 57 58 59
16 48 17 49 50 19 51 53 55
32 33 34 35 37 38 39 43 47
1 D 0 for i < n. But from Denition 11.9 of the order of (see so i page 852), n is the smallest integer such that n D 1, hence we get a contradiction. 2. Let 1 ; 2 ; : : : ; L be elements of some extension eld of G F.q/, and let `1 ; `2 ; : : : ; ` L be the orders of these elements, respectively. Moreover, let m 1 .x/, m 2 .x/; : : : , m L .x/ be the minimum functions of these elements with coefcients from G F.q/, and let g.x/ be the smallest monic polynomial with coefcients from G F.q/ that has 1 ; 2 ; : : : ; L as roots: then a) g.x/ D l:c:m:.m 1 .x/; m 2 .x/; : : : ; m L .x//; b) if the minimum functions are all distinct, that is they do not have factor polynomials in common, then g.x/ D m 1 .x/ m 2 .x/ : : : m L .x/; c) if n D l:c:m:.`1 ; `2 ; : : : ; ` L /, then x n for i < n. 1 D h.x/ g.x/, and x i 1 6D h.x/ g.x/
862
Proof. a) Noting that g.x/ must be divisible by each of the minimum functions, it must be the smallest degree monic polynomial divisible by m 1 .x/; m 2 .x/; : : : , m L .x/, but this is just the denition of the least common multiple. b) If all the minimum functions are distinct, as each is irreducible, the least common multiple is given by the product of the polynomials. c) As n is a multiple of the order of each element, n 1 D 0, for j D 1; 2; : : : ; L; j then x n 1 must be divisible by m j .x/, for j D 1; 2; : : : ; L, and therefore it must be divisible by the least common multiple of these polynomials. Assume now that g.x/ divides x i 1 for i < n; then ij 1 D 0 for each j D 1; 2; : : : ; L, and thus i is a multiple of `1 ; `2 ; : : : ; ` L . But n is the smallest integer multiple of `1 ; `2 ; : : : ; ` L , hence we get a contradiction. We note that if the extension eld is G F.q k / and L D q k and h.x/ D 1. 1 D n, then g.x/ D x n 1
11.2.3
Cyclic codes
In Section 11.2.1 we dealt with the theory of binary group codes. We now discuss a special class of linear codes. These codes, called cyclic codes, are based upon polynomial algebra and lead to particularly efcient implementations for encoding and decoding.
Many such factorizations are possible for a given polynomial x n 1; we will consider any one of them. We denote the degrees of g.x/ and h.x/ as r and k, respectively; thus n D k C r. The choice of the symbols n, k and r is intentional, as they assume the same meaning as in the previous sections. The polynomial arithmetic modulo q.x/ D x n 1 is particularly important in the discussion of cyclic codes. Proposition 11.2 Consider the set of all polynomials of the form c.x/ D a.x/ g.x/ modulo q.x/, as a.x/ ranges over all polynomials of all degrees with coefcients from G F.q/. This set must be nite as there are at most q n remainder polynomials that can be obtained by dividing a polynomial by x n 1. Now we show that there are exactly q k distinct polynomials. Proof. There are at least q k distinct polynomials a.x/ of degree less than or equal to k 1, and each such polynomial leads to a distinct polynomial a.x/ g.x/. In fact, as the degree
863
of a.x/ g.x/ is less than r C k D n, no reduction modulo x n 1 is necessary for these polynomials. Now let a.x/ be a polynomial of degree greater than or equal to k. To reduce the polynomial a.x/ g.x/ modulo x n 1, we divide by x n 1 and keep the remainder; thus a.x/ g.x/ D Q.x/ .x n 1/ C r.x/ (11.97)
where 0 deg.r.x// < n. By using (11.96), we can express r.x/ as r.x/ D .a.x/ h.x/ Q.x//g.x/ D a 0 .x/ g.x/ (11.98)
As r.x/ is of degree less than n, a 0 .x/ is of degree less than k, but we have already considered all polynomials of this form; therefore r.x/ is one of the q k polynomials determined in the rst part of the proof. Example 11.2.22 Let g.x/ D x C 1, G F.q/ D G F.2/, and n D 4; then all polynomials a.x/ g.x/ modulo x 4 1 D x 4 C 1 are given by a.x/ 0 1 x x C1 x2 2C1 x x2 C x x2 C x C 1 a.x/ g.x/ mod .x 4 0 x C1 x2 C x x2 C 1 x3 C x2 3 C x2 C x C 1 x x3 C x x3 C 1 1/ code word 0000 1100 0110 1010 0011 1111 0101 1001
(11.99)
We associate with any polynomial of degree less than n and coefcients from G F.q/ a vector of length n with components equal to the coefcients of the polynomial, that is f .x/ D f 0 C f 1 x C f 2 x 2 C C f n
1x n 1
1/
(11.100)
Note that in the denition f n 1 does not need to be non-zero. We can now dene cyclic codes. The code words will be the vectors associated with a set of polynomials; alternatively, we speak of the polynomials themselves as being code words or code polynomials (see (11.99)). Denition 11.12 Choose a eld G F.q/, a positive integer n and a polynomial g.x/ with coefcients from G F.q/ such that x n 1 D g.x/ h.x/; furthermore, let deg.g.x// D r D n k. Words of a cyclic code are the vectors of length n that are associated with all multiples of g.x/ reduced modulo x n 1. In formulae: c.x/ D a.x/ g.x/ mod .x n 1/, for a.x/ polynomial with coefcients from G F.q/. The polynomial g.x/ is called a generator polynomial.
864
(11.102)
Proof. If c.x/ is a code polynomial, then c.x/ D a.x/g .x/ mod.x n 1/, but h.x/ c.x/ D h.x/ a.x/ g.x/ D a.x/ .g.x/h.x// D a.x/.x n 1/ D 0 mod.x n 1/. Assume now h.x/ c.x/ D 0 mod.x n 1/; then h.x/ c.x/ D Q.x/.x n 1/ D Q.x/ h.x/ g.x/, or c.x/ D Q.x/ g.x/, therefore c.x/ is a code polynomial. 5. Let x n 1 D g.x/ h.x/, where g.x/ D g0 C g1 x C C gr x r and h.x/ D h 0 C h 1 x C 1 has the C h k x k ; then the code corresponding to all multiples of g.x/ modulo x n
865
(11.103)
3 : : : h1 h0 : : : h0 0 7 7 5 ::: 0 0
(11.104)
Proof. We show that G is the generator matrix. The rst row of G corresponds to the polynomial g.x/, the second to xg.x/ and the last row to x k 1 g.x/, but the code words are all words of the form g.x// (11.105) But (11.105) expresses all code words as linear combinations of the rows of G, therefore G is the generator matrix of the code. To show that H is the parity check matrix, we consider the product c.x/ h.x/. If we write c.x/ D c0 C c1 x C C cn and h.x/ D h 0 C h 1 x C C h k 1 x k 1 C h k x k C C h n below this point where h kC1 D h kC2 D D h n 1 D 0, we get d.x/ D c.x/ h.x/ D d0 C d1 x C C d2n where 8 i >X > > c j hi > < > > > > :
jD0 n 1 X j 2x 1x n 1 1x n 1
.a0 C a1 x C C ak
1x
k 1
1 .x
k 1
(11.106)
(11.107)
2n 2
(11.108)
if 0 i n if n i 2n
1 (11.109) 2
di D
c j hi
O O O 1, and denote the result as d.x/ D d0 C d1 x C We consider reducing d.x/ O O C dn 1 x n 1 ; then di D di C dnCi , i D 0; 1; 2; : : : ; n 1. If c.x/ h.x/ D 0 mod.x n 1/, O then di D 0, i D 0; 1; 2; : : : ; n 1, therefore we get
i X jD0
jDi .n 1/ modulo x n
c j hi
n 1 X jDi C1
c j h nCi
D0
i D 0; 1; 2; : : : ; n
(11.110)
866
For i D n
1, (11.110) becomes
n 1 X jD0
c j hn
1] T
1 j
D0
(11.111)
or [h n 1 h n For i D n
D 0.
n 2 X jD0
c j hn
2 j
C cn
1 hn 1
D0
(11.112)
n r X jD0
c j hn hn
n 1 X jDn r C1
c j h 2n
1] T
D0
(11.113)
or [h n r h n r 1 : : : h n in matrix form as 2 6 6 6 4
r C2
r C1 ] [c0 c1
: : : cn
0 h n r h n r 1 : : : h n r C2 h n r C1 cn 1 therefore all code words are solutions of the equation Hc D 0, where H is given by (11.104). It still remains to be shown that all solutions of the equation Hc D 0 are code words. As h n 1 D h n 2 D D h n r C1 D 0, and h 0 6D 0, from (11.104) H has rank r, and can be written as H D [A B], where B is an r r matrix with non-zero determinant; therefore 2 3 2 3
hn hn : : :
1 2
hn hn
2 3
::: :::
h1 h0
h0 hn 1 : : :
32 76 76 76 54
c0 c1 : : :
7 6 7 6 7D6 5 4
0 0 : : :
3 7 7 7 5 (11.114)
7 6 ckC1 7 6 7 7 6 6 (11.115) 7 6 : 7 D B 1A6 5 4 : 5 4 : cn 1 ck 1 so there are q k D q n r solutions of the equation Hc D 0. As there are q k code words, all solutions of the equation Hc D 0 are the code words in the cyclic code. Example 11.2.24 Let q D 2 and n D 7. As x 7 1 D x 7 C 1 D .x 3 C x C 1/.x 3 C x 2 C 1/.x C 1/, we can choose g.x/ D x 3 C x C 1 and h.x/ D .x 3 C x 2 C 1/.x C 1/ D x 4 C x 2 C x C 1; thus the
ck
c0 c1 : : :
867
by 1 0 1 0 0 0 3 (11.116) 7 1 1 0 1 0 07 7 0 1 1 0 1 07 5 0 0 1 1 0 1 3 0 1 0 1 1 1 1 0 1 1 1 05 0 1 1 1 0 0
(11.117)
Note that the columns of H are all possible non-zero vectors of length 3, so the code is a Hamming single error correcting (7,4) code. 6. In a code word, any string of r consecutive symbols, even taken cyclically, can identify the check positions. Proof. From (11.115) it follows that the last r positions can be check positions. Now, if we cyclically permute every code word of m positions, the resultant words are themselves code words; thus the r check positions can be cyclically permuted anywhere in the code words. 7. As the r check positions can be the rst r positions, a simple encoding method in canonical form is given by the following steps. Step 1: represent the k information bits by the coefcients of the polynomial m.x/ D m0 C m1 x C C mk 1 x k 1. Step 2: multiply m.x/ by x r to obtain x r m.x/. Step 3: divide x r m.x/ by g.x/ to obtain the remainder r.x/ D r0 C r1 x C C rr Step 4: form the code word c.x/ D .x r m.x/ are the parity check bits.
1
xr
1.
Proof. To show that .x r m.x/ r.x// is a code word, we must prove that it is a multiple of g.x/: from Step 3 we obtain x r m.x/ D Q.x/ g.x/ C r.x/ so that .x r m.x/ r.x// D Q.x/ g.x/ (11.119) (11.118)
Example 11.2.25 Let g.x/ D 1 C x C x 3 , for q D 2 and n D 7. We report in Table 11.12 the message words .m 0 ; : : : ; m 3 / and the corresponding code words .c0 ; : : : ; c6 / obtained by the generator polynomial according to Denition 11.12 on page 863 for a.x/ D m.x/; the same code in canonical form, obtained by (11.119), is reported in Table 11.13.
868
Message .m 0 m 1 m 2 m 3 / 0000 1000 0100 1100 0010 1010 0110 1110 0001 1001 0101 1101 0011 1011 0111 1111
Code .c0 c1 c2 c3 c4 c5 c6 / 0000000 1101000 0110100 1011100 0011010 1110010 0101110 1000110 0001101 1100101 0111001 1010001 0010111 1111111 0100011 1001011
0g.x/ D 0 1g.x/ D 1 C x C x 3 xg.x/ D x C x 2 C x 4 .1 C x/g.x/ D 1 C x 2 C x 3 C x 4 x 2 g.x/ D x 2 C x 3 C x 5 .1 C x 2 /g.x/ D 1 C x C x 2 C x 5 .x C x 2 /g.x/ D x C x 3 C x 4 C x 5 .1 C x C x 2 /g.x/ D 1 C x 4 C x 5 x 3 g.x/ D x 3 C x 4 C x 6 .1 C x 3 /g.x/ D 1 C x C x 4 C x 6 .x C x 3 /g.x/ D x C x 2 C x 3 C x 6 .1 C x C x 3 /g.x/ D 1 C x 2 C x 6 .x 2 C x 3 /g.x/ D x 2 C x 4 C x 5 C x 6 2 C x 3 /g.x/ D 1 C x C x 2 C x 3 C x 4 C x 5 C x 6 .1 C x .x C x 2 C x 3 /g.x/ D x C x 5 C x 6 .1 C x C x 2 C x 3 /g.x/ D 1 C x 3 C x 5 C x 6
Table 11.13 (7,4) binary cyclic code in canonical form, generated by g.x/ D 1 C x C x3 .
Message .m 0 m 1 m 2 m 3 / 0000 1000 0100 1100 0010 1010 0110 1110 0001 1001 0101 1101 0011 1011 0111 1111
Code .c0 c1 c2 c3 c4 c5 c6 / 0000000 1101000 0110100 1011100 1110010 0011010 1000110 0101110 1010001 0111001 1100101 0001101 0100011 1001011 0010111 1111111
869
C C g1 x C g0
(11.120)
gr 1 .gr
1x
r 1
C gr
2x
r 2
C C g1 x C g0 / mod g.x/
(11.121)
C gr
2x
r 2
C C g1 x C g0 / mod g.x/
(11.122)
is the remainder after dividing m 0 x r by g.x/. We now consider the scheme illustrated in Figure 11.3, where multiplications and additions are in G F.q/, and Tc denotes the clock period with which the message symbols fm i g, i D k 1; : : : ; 1; 0, are input to the shift register. In the binary case, the storage elements are ip ops, the addition is the modulo 2 addition, and multiplication by gi is performed by a switch that is open or closed depending upon whether gi D 0 or 1, respectively. Note that if m 0 is input, the storage elements of the shift register will contain the coefcients of the remainder upon dividing m 0 x r by g.x/. Let us suppose we want to compute the remainder upon dividing m 1 x r C1 by g.x/. We could rst compute the remainder of the division of m 1 x r by g.x/, by presenting m 1 at the input, then multiplying the remainder by x, and again reduce the result modulo g.x/. But once the remainder of the rst division is stored in the shift register, multiplication by x and division by g.x/ are obtained simply by clocking the register once with no input. In fact, if the shift register contains the polynomial b.x/ D b0 C b1 x C C br
1x r 1
(11.123)
g0 Tc
g r1 Tc Tc IN OUT mi
1 gr
Figure 11.3. Scheme of an encoder for cyclic codes using a shift register with r elements.
870
gr 1 .gr
1
1x
r 1
C C g1 x C g0 // mod g.x/
1 gr 1
(11.124)
g0 C .b0 br
1 gr 1
br gr
g1 / x C mod g.x/
1/ x
r 1
that is just the result obtained by clocking the register once. Finally, we note that superimposition holds in computing remainders; in other words, if m 0 x r D r1 .x/ mod g.x/ and m 1 x r C1 D r2 .x/ mod g.x/, then m 0 x r C m 1 x r C1 D r1 .x/Cr2 .x/ mod g.x/. Therefore, to compute the remainder upon dividing m 0 x r Cm 1 x r C1 by g.x/ using the scheme of Figure 11.3, we would rst input m 1 and then next input m 0 to the shift register. Hence, to compute the remainder upon dividing x r m.x/ D m 0 x r C m 1 x r C1 C C m k 1 x n 1 by g.x/ we input the symbols m k 1 ; m k 2 ; : : : ; m 1 ; m 0 to the device of Figure 11.3; after the last symbol, m 0 , enters, the coefcients of the desired remainder will be contained in the storage elements. From (11.119) we note that the parity check bits are the inverse elements (with respect to addition) of the values contained in the register. In general for an input z.x/, polynomial with n coefcients, after n clock pulses the device of Figure 11.3 yields x r z.x/ mod g.x/.
C hk
1 m0
1 m0
C C h1 mk
2 m1
C h0 mk
2
D0
1/
(11.125) (11.126)
h k 1 .h k h k 1 .h k
C hk
C C h1 mk
C h0 mk
pr
C hk
2 m0
C C h1 mk
C h0 mk
2/
(11.127)
and so forth. Let us consider the scheme of Figure 11.4 and assume that the register initially contains the symbols m 0 ; m 1 ; : : : ; m k 1 . After one clock pulse m k 1 will appear at the output, all information symbols will have moved by one place to the right and the parity check symbol pr 1 will appear in the rst left-most storage element; after the second clock pulse, m k 2 will appear at the output, all symbols contained in the storage elements will move one place to the right and the parity check symbol pr 2 will appear in the left-most storage element. It is easy to verify that, if we apply n clock pulses to the device, the output will be given by the k message symbols followed by the r parity check bits.
871
hk1
h k1 Tc
h k2 Tc
h k3 Tc
h1 Tc
h0
OUT
IN
Figure 11.4. Scheme of an encoder for cyclic codes using a shift register with k elements.
As upon dividing a j1 g.x/ and a j2 g.x/ by g.x/ we get 0 as a remainder, the division of z 1 .x/ and z 2 .x/ by g.x/ gives the same remainder, namely the polynomial ri .x/, where i .x/ D Q.x/ g.x/ C ri .x/ deg.ri .x// < deg.g.x// D r (11.130)
Now assume z 1 .x/ and z 2 .x/ are in different cosets, say, the i 1 -th and i 2 -th cosets, but have the same remainder, say, r0 .x/, if they are divided by g.x/; then the coset leaders i1 .x/ and i2 .x/ of these cosets must give the same remainder r0 .x/ if they are divided by g.x/, i.e. i1 .x/ D Q 1 .x/ g.x/ C r0 .x/ (11.131)
872
g0 IN z 0 z 1 ... z n2 z n1 Tc
g r1 Tc Tc
1 gr
Figure 11.5. Device to compute the division of the polynomial z.x/ D z0 Cz1 xC Czn 1 xn 1 by g.x/. After n clock pulses the r storage elements contain the remainder r0 ; r1 ; : : : ; rr 1 .
and i2 .x/ D Q 2 .x/ g.x/ C r0 .x/ therefore we get i2 .x/ D i1 .x/ C .Q 2 .x/ Q 1 .x//g.x/ D i1 .x/ C Q 3 .x/ g.x/ (11.133) (11.132)
This implies that i1 .x/ and i2 .x/ are in the same coset, which is a contradiction. This result leads to the following decoding method for cyclic codes. Step 1: compute the remainder upon dividing the received polynomial z.x/ of degree n 1 by g.x/, for example, by the device of Figure 11.5 (see (11.124)), by presenting at the input the sequence of received symbols, and applying n clock pulses. The remainder identies the coset leader of the coset where the received polynomial is located. Step 2: subtract the coset leader from the received polynomial to obtain the decoded code word.
Hamming codes
Hamming codes are binary cyclic single error correcting codes. We consider cyclic codes over G F.2/, where g.x/ is an irreducible polynomial of degree r such that g.x/ divides r x 2 1 1, but not x ` 1 for ` < 2r 1. To show that g.x/ is a primitive irreducible polynomial, we choose n D 2r 1, thus n 1 D g.x/ h.x/ and the corresponding cyclic code has parameters n D 2r 1, r, and x k D 2r 1 r. Proposition 11.4 H This code has minimum distance dmin D 3 and therefore is a single error correcting code.
H Proof. We rst prove that dmin 3 by showing that all single error polynomials have distinct, non-zero remainders if they are divided by g.x/.
873
Assume that x i D 0 mod g.x/, for some 0 i n 1; then x i D Q.x/ g.x/, which is impossible since g.x/ is not divisible by x. Now assume that x i and x j give the same remainder upon division by g.x/, and that 0 i < j n 1; then xj x i D x i .x j
i
1/ D Q.x/ g.x/
(11.134)
but g.x/ does not divide x i , so it must divide .x j i 1/. But 0 < j i n 1 and by H assumption g.x/ does not divide this polynomial. Hence dmin 3. By the limit (11.15) we know that for a code with xed n and k the following inequality holds: 2 3 n n n n (11.135) C C C jdH 2k 41 C 1k 5 2 min 1 2 2 As n D 2r 2 41 C 1 and k D n r 2 1 1 r, we have r 2 2 1 2r C C jdH 3 1 r 1k 5 2
min
(11.136)
2 but 1C
H and therefore dmin 3.
r 2 1
D 2r
(11.137)
We have seen in the previous section how to implement an encoder for a cyclic code. We consider now the decoder device of Figure 11.6, whose operations are described as follows. 1. Initially all storage elements of the register contain zeros and the switch SW is in position 0. The received n-bit word z D .z 0 ; : : : ; z n 1 / is sequentially clocked into the lower register, with n storage elements, and into the feedback register, with r storage elements, whose content is denoted by r0 ; r1 ; : : : ; rr 1 . 2. After n clock pulses, the behavior of the decoder depends on the value of v: if v D 0, the switch SW remains in the position 0 and both registers are clocked once. This procedure is repeated until v D 1, which occurs for r0 D r1 D D rr 2 D 0; then SW moves to position 1 and the content of the last stage of the feedback shift register is added modulo 2 to the content of the last stage of the lower register; both registers are then clocked until the n bits of the entire word are obtained at the output of the decoder. Overall, 2n clock pulses are needed.
874
g x0 Tc
g2 x1 Tc NOR v Tc
g r1 x r2 x r1 Tc
0 (v=0) SW 1 (v=1)
z IN x0 Tc x1 Tc x n1 Tc OUT
Figure 11.6. Scheme of a decoder for binary cyclic single error correcting codes (Hamming codes). All operations are in GF.2/.
^ c
We now illustrate the procedure of the scheme of Figure 11.6. First of all we note that for the rst n clocks the device coincides with that of Figure 11.3, hence the content of the shift register is given by (11.138) r.x/ D x r z.x/ mod g.x/ We consider two cases. 1. The received word is correct, z.x/ D c.x/. After the rst n clock pulses, from (11.138) we have (11.139) r.x/ D x r c.x/ D x r a.x/ g.x/ D 0 mod g.x/ and thus v D 1 and rr In the successive n clock pulses we have ci D z i C 0 O O therefore c D c. 2. The received word is affected by one error, z.x/ D c.x/ C x i . In other words we assume that there is a single error in the i-th bit, 0 i n 1. After the rst n clock pulses, it is r.x/ D x r x i mod g.x/ (11.142) i D 0; : : : ; n 1 (11.141)
1
D0
(11.140)
875
If i D n
1, we have r.x/ D x r x n D xn xr D .x n
1 1
1/ x r
C xr
1
1 1
C xr
D 1 and rr
D D r0 D 0 .v D 1/
(11.144)
This leads to switching SW , therefore during the last n clock pulses we have cn O
1
D zn
C1
ci D z i C 0 O
i Dn
2; : : : ; 0
(11.145)
Therefore the bit in the last stage of the buffer is corrected. If i D n j, then we have r.x/ D x r j
(11.146)
thus only at the (n C j 1)-th clock pulse the condition (11.144) that forces to switch SW from 0 to 1 occurs; therefore, at the next clock pulse the received bit in error will be corrected.
; 0; : : : ; 0/
(11.147)
where within the two 1s the values can be either 0 or 1. Then we can write the vector e in polynomial form, e.x/ D x j B.x/ (11.148)
where B.x/ is a polynomial of degree ` 1 n k 1. Thus e.x/ is divisible by the generator polynomial g.x/ if B.x/ is divisible by g.x/, as x is not a factor of g.x/; but B.x/ has a degree at most equal to .n k 1/, lower than the degree of g.x/, equal to n k; therefore e.x/ cannot be a code word. We have then that all burst errors of length ` less than or equal to r D n k are detectable by .n; k/ cyclic codes. This result leads to the introduction of the cyclic redundancy check (CRC) codes.
11.2.4
We consider a class of cyclic codes over G F.q/ such that the Hamming distance between every pair of distinct code words is a constant; this is equivalent to stating that the weight
876
n 7 15 31 63 127
k 3 4 5 6 7
r 4 11 26 57 120
dmin 4 8 16 32 64
of all non-zero code words is equal to the same constant. We show that in the binary case, for these codes the non-zero code words are related to the PN sequences of Appendix 3.A. Let n D q k 1, and x n 1 D g.x/ h.x/, where we choose h.x/ as a primitive polynomial of degree k; then the resultant code has minimum distance
H dmin D .q
1/ q k
(11.149)
The parameters of some binary codes in this class are listed in Table 11.14. To show that these codes have minimum distance given by (11.149), rst we prove the following: Property. All non-zero code words have the same weight. Proof. We begin by showing that x i g.x/ 6D x j g.x/ mod.x n 1/ 0i < j n 1 (11.150)
1/ D Q.x/ h.x/
(11.152)
But this is impossible since h.x/ is a primitive polynomial of degree k and cannot divide .x j i 1/, as . j i/ < n D .q k 1/. Relation (11.150) implies that all cyclic shifts of the code polynomial g.x/ are unique, but there are n D .q k 1/ cyclic shifts. Furthermore we know that there are only q k code words and one is the all-zero word; therefore all cyclic shifts of g.x/ are all the non-zero code words and they all have the same weight. Recall Property 2 of a group code (see page 832), that is if all code words of a linear code are written as rows of a matrix, every column is either formed by all zeros, or it consists of each eld element repeated an equal number of times. If we apply this result to
877
a simplex code, we nd that no column can be all zero as the code is cyclic, so the sum of the weights of all code words is given by sum of weights D n.q 1/ qk D .q k q 1/ .q 1/ q k
1
(11.153)
But there are .q k 1/ non-zero code words, all of the same weight; the weight of each word is then given by weight of non-zero code words D .q
H dmin D .q
1/ q k
(11.154)
(11.155)
Example 11.2.26 H Let q D 2, n D 15, and k D 4; hence r D 11, and dmin D 8. Choose h.x/ as a primitive irreducible polynomial of degree 4 over G F.2/, h.x/ D x 4 C x C 1. The generator polynomial g.x/ is obtained by dividing x 15 1 by h.x/ D x 4 C x C 1 in G F.2/, obtaining g.x/ D x 11 C x 8 C x 7 C x 5 C x 3 C x 2 C x C 1 (11.156)
Given an extension eld G F.2k / and n D 2k 1, from Property 2 on page 861, x n 1 is given by the l.c.m. of the minimum functions of the elements of the extension eld. As h.x/ is a primitive polynomial, g.x/ is therefore given by the l.c.m. of the minimum functions of the elements 1; 3 ; 5 ; 7 , from G F.24 /. By a table similar to Table 11.11, obtained for G F.26 /, and using one of the three methods to determine the minimum function (see page 859), it turns out that the generator polynomial for this code is given by g.x/ D .x C 1/.x 4 C x 3 C x 2 C x C 1/.x 2 C x C 1/.x 4 C x 3 C 1/ (11.157)
Relation to PN sequences
We consider a periodic binary sequence of period L, given by : : : ; p. 1/; p.0/, p.1/, : : : , with p.`/ 2 f0; 1g. We dene the normalized autocorrelation function of this sequence as # " L 1 X 1 r p .m/ D . p.`/ p.` m// (11.158) L 2 L `D0 Note that with respect to (3.302), now p.`/ 2 f0; 1g rather than p.`/ 2 f 1; 1g. Theorem 11.1 If the periodic binary sequence f p.`/g is formed by repeating any non-zero code word of a simplex binary code of length L D n D 2k 1, then 8 m D 0; L ; 2L ; : : : <1 (11.159) r p .m/ D : 1 otherwise L
878
Proof. We recall that for a simplex binary code all non-zero code words a) have weight 2k
1,
b) are cyclic permutations of the same code word. As the code is linear, the Hamming distance between any code word and a cyclic permutation of this word is 2k 1 ; this means that for the periodic sequence formed by repeating any non-zero code word we obtain ( L 1 X 0 m D 0; L ; 2L ; : : : . p.`/ p.` m// D (11.160) 2k 1 otherwise `D0 Substitution of (11.160) in (11.158) yields 8 >1 < r p .m/ D 2k 1 2k 1 > : D k k 2 1 2 1
m D 0; L ; 2L ; : : : (11.161) otherwise
If we recall the implementation of Figure 11.4, we nd that the generation of such sequences is easy. We just need to determine the shift register associated with h.x/, load it with anything except all zeros, and let it run. For example, choosing h.x/ D x 4 C x C 1, we get the PN sequence of Figure 3.41, as illustrated in Figure 11.7, where L D n D 24 1 D 15.
11.2.5
BCH codes
879
p(l)
p(l1)
p(l2)
p(l3)
p(l4)
l 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
p(l1) 0 1 0 0 1 1 0 1 0 1 1 1 1 0 0 0
p(l2) 0 0 1 0 0 1 1 0 1 0 1 1 1 1 0 0
p(l3) 0 0 0 1 0 0 1 1 0 1 0 1 1 1 1 0
p(l4) 1 0 0 0 1 0 0 1 1 0 1 0 1 1 1 1
Figure 11.7. Generation of a PN sequence as a repetition of a code word of a simplex code with L D n D 15.
code words. The form of (11.162) resembles equation (11.20), where H is the generalized parity check matrix. One obvious difference is that in (11.20) H and c have elements from the same eld, whereas this does not occur for the vector equation (11.162). However, this difference is not crucial as each element from G F.q m / can be written as a vector of length m with elements from G F.q/. Thus each element .i / j in the matrix is replaced by a column vector with m components. The resultant matrix, with Lm rows and n columns, consists of elements from G F.q/ and is therefore just a generalized parity check matrix for the considered code. From the above discussion it appears that, if L roots are specied, the resultant linear code has r D Lm parity check symbols, as the parity check matrix has r D Lm rows. However, not all rows of the matrix are necessarily independent; therefore the actual number of parity check symbols may be less than Lm. We now show that if n is properly chosen, the resultant codes are cyclic codes. Let m j .x/ be the minimum function of j , j D 1; 2; : : : ; L, where j 2 G F.q m / and m j .x/ has coefcients from G F.q/. For Property 3 on page 858, every code polynomial c.x/ must be divisible by m 1 .x/; m 2 .x/; : : : ; m L .x/, and is thus divisible by the least common multiple of such minimum functions, l:c:m:.m 1 .x/; m 2 .x/; : : : ; m L .x//. If we dene g.x/ D l:c:m:.m 1 .x/; m 2 .x/; : : : ; m L .x// (11.163)
880
then all multiples of g.x/ are code words. In particular from Denition 11.12 the code is cyclic if (11.164) x n 1 D g.x/ h.x/ Let `i be the order of i , i D 1; 2; : : : ; L, and furthermore let n D l:c:m:.`1 ; `2 ; : : : ; ` L / (11.165)
From the properties of the minimum function (see Property 2, page 861), we know that g.x/ divides x n 1; thus the code is cyclic if n is chosen as indicated by (11.165). We note that r D deg.g.x// m L (11.166)
as deg.m i .x// m. We see that r is equal to m L if all minimum functions are distinct and are of degree m; conversely, r < m L if any minimum function has degree less than m or if two or more minimum functions are identical. Example 11.2.27 Choose q D 2 and let be a primitive element of G F.24 /; furthermore let the code polynomials have as roots the elements , 2 , 3 , 4 . To derive the minimum functions of the chosen elements we look up for example the Appendix C of [3], where such functions are listed. Minimum functions and orders of elements chosen for this example are given in Table 11.15. Then g.x/ D .x 4 C x C 1/ .x 4 C x 3 C x 2 C x C 1/ (11.167) n D l:c:m:.15; 15; 5; 15/ D 15
H The resultant code is therefore a (15,7) code; later we will show that dmin D 5.
Roots 2 3 4
Minimum function x4 C x C 1 x4 C x C 1 4 C x3 C x2 C x C 1 x x4 C x C 1
Order 15 15 5 15
881
The basic mathematical fact required to prove the error correcting capability of BCH codes is that if 1 ; 2 ; : : : ; r are elements from any eld, the determinant of the Vandermonde matrix, given by 1 1 ::: 1 1 2 : : : r 2 2 2 2 : : : r (11.168) det 1 : : : : : : r 1 r 1 : : : r 1
1 2 r
is non-zero if and only if i 6D j , for all indices i 6D j. In particular, we prove the following result. Lemma. The determinant (11.168) is given by 1 1 ::: 1 1 2 : : : r 2 r .r C1/ 2 2 2 : : : r D . 1/ 2 D D det 1 : : : : : : r 1 r 1 : : : r 1 r 1 2 as 1 2 2 2
r 2 1
r Y
i; j D 1 i< j
.i
j/
(11.169)
1 r 2 r : : :
1
r : : : r
(11.170)
so that D D P.1 /. Now, P.x/ is a polynomial of degree at most r 1 whose zeros are x D 2 ; x D 3 ; : : : ; x D r , because if x D i , i D 2; 3; : : : ; r, the determinant D is equal to zero as two columns of the matrix are identical. Thus P.x/ D k1 .x and D D P.1 / D k1 .1 2 /.1 3 / : : : .1 r / (11.172) It remains to calculate k1 . The constant k1 is the coefcient of x r 1 ; therefore from (11.170) we get 1 1 ::: 1 2 3 : : : r r . 1/ k1 D det : : D k2 .2 3 /.2 4 / : : : .2 r / (11.173) : : : : r 2 r 2 : : : r 1
2 3 r
2 /.x
3 / : : : .x
r /
(11.171)
882
k2 D k3 .3
4 /.3
5 / .3
r / r / (11.174)
.i
j / D . 1/
r .r C1/ 2
r Y
i; j D 1 i< j
.i
j/
(11.175)
Theorem 11.2 Consider a code with symbols from G F.q/, whose code polynomials have as zeros the elements m 0 ; m 0 C1 ; : : : ; m 0 Cd 2 , where is any element from G F.q m / and m 0 is any integer. Then the resultant .n; k/ cyclic code has the following properties:
H a) it has minimum distance dmin d if the elements m 0 ; m 0 C1 ; : : : ; m 0 Cd distinct; m l b) n k .d 1/m; if q D 2 and m 0 D 1, then n k d 2 1 m; 2,
are
c) n is equal to the order of , unless d D 2, in which case n is equal to the order of m 0 ; d) g.x/ is equal to the least common multiple of the minimum functions of m 0 , m 0 C1 , : : : , m 0 Cd 2 . Proof. The proof of part d) has already been given (see (11.163)); the proof of part b) then follows by noting that each minimum function is at most of degree m, and there are at most .d 1/ distinct minimum functions. If q D 2 and m 0 D 1, the minimum function of the m raised to an even power, for example 2i , is the same as l minimum function of i d 1 m distinct minimum (see Property 2 on page 859), therefore there are at most 2 functions. To prove part c) note that, if d D 2, we have only the root m 0 , so that n is equal to the order of m 0 . If there is more than one root, then n must be the least common multiple of the order of the roots. If m 0 and m 0 C1 are both roots, then . m 0 /n D 1 and . m 0 C1 /n D 1, so that n D 1; thus n is a multiple of the order of . On the other hand, if ` is the order of , . m 0 Ci /` D . ` /m 0 Ci D 1m 0 Ci D 1; therefore ` is a multiple of the order of every root. Then n is the least common multiple of numbers all of which divide `, and therefore n `; thus n D `.
883
Finally we prove part a). We note that the code words must satisfy the condition 3 2 3 2 2 3 c0 . m 0 /2 ::: . m 0 /n 1 1 m0 0 6 c1 7 6 1 m 0 C1 7 607 . m 0 C1 /2 : : : . m 0 C1 /n 1 7 6 7 6 c2 7 6 7 6 (11.176) 76 6 : 7D6 : 7 : : : 56 : 7 4 : 5 4 : : : 4 : 5 : 0 1 m 0 Cd 2 . m 0 Cd 2 /2 : : : . m 0 Cd 2 /n 1 cn 1 We now show that no linear combination of .d 1/ or fewer columns is equal to 0. We do this by showing that the determinant of any set of .d 1/ columns is non-zero. Choose columns j1 ; j2 ; : : : ; jd 1 ; then . m 0 / j1 . m 0 / j2 ::: . m 0 / jd 1 . m 0 C1 / j1 . m 0 C1 / j2 : : : . m 0 C1 / jd 1 (11.177) det : : : : : : . m 0 Cd 2 / j1 . m 0 Cd 2 / j2 : : : . m 0 Cd 2 / jd 1 1 1 ::: 1 j1 j2 ::: jd 1 D m 0 . j1 C j2 CC jd 1 / det (11.178) : : : : : : . j1 /d 2 . j2 /d 2 : : : . jd 1 /d 2 D m 0 . j1 C j2 CC jd
1/
. 1/
.d 1/d 2
d 1 Y
i; k D 1 i <k
. ji
jk / 6D 0
(11.179)
Note that we have proven that .d 1/ columns of H are linearly independent even if they are multiplied by elements from G F.q m /. All that would have been required was to show linear independence if the multipliers are from G F.q/.
(11.181)
884
Roots 1 2 4 8 16 32 3 6 12 24 48 33 5 10 20 40 17 34 7 14 28 56 49 35 9 18 36 11 22 44 25 50 37 13 26 52 41 19 38 15 30 60 57 51 39 21 42 23 46 29 58 53 43
Minimum function x6 C x C 1 x6 C x4 C x2 C x C 1 x6 C x5 C x2 C x C 1 x6 C x3 C 1 x3 C x2 C 1 x6 C x5 C x3 C x2 C 1 x6 C x4 C x3 C x C 1 x6 C x5 C x4 C x2 C 1 x2 C x C 1 x6 C x5 C x4 C x C 1 x3 C x C 1 x6 C x5 C 1
27 54 45 31 62 61 59 55 47
Example 11.2.28 Consider binary BCH codes of length 63, that is q D 2 and m D 6. To get a code with design distance d we choose as roots ; 2 ; 3 ; : : : ; d 1 , where is a primitive element from G F.26 /. Using Table 11.11 on page 861, we get the minimum functions of the elements from G F.26 / given in Table 11.16. Then the roots and generator polynomials for different values of d are given in Table 11.17; the parameters of the relative codes are given in Table 11.18. Example 11.2.29 Let q D 2, m D 6, and choose as roots , 2 , 3 , 4 , with D 3 , where is a primitive m H element of G F.26 /; then n D 2 c 1 D 63 D 21, dmin d D 5, and 3 g.x/ D l.c.m.m 3 .x/; m 6 .x/; m 9 .x/; m 12 .x// D m 3 .x/m 9 .x/ D .x C x C x C x C 1/.x C x C 1/
6 4 2 3 2
(11.183)
As r D deg.g.x// D 9, then k D n
Example 11.2.30 Let q D 2, m D 4, and choose as roots , 2 , 3 , 4 , with primitive element of G F.24 /; H then a (15,5) code is obtained having dmin 5, and g.x/ D .x 4 Cx C1/.x 4 Cx 3 Cx 2 Cx C1/.
885
Table 11.17 Roots and generator polynomials of BCH codes of length n D 63 D 26 1 for different values of d. is a primitive element of GF.26 / (see (11.180)).
d 3 5 7 9 11 13 15 21 23 27 31
Roots 2 2 3 4 2 : : : 6 2 : : : 8 2 : : : 10 2 : : : 12 2 : : : 14 2 : : : 20 2 : : : 22 2 : : : 26 2 : : : 30
Generator polynomial .x 6 C x C 1/ D g3 .x/ .x 6 C x C 1/.x 6 C x 4 C x 2 C x C 1/ D g5 .x/ .x 6 C x 5 C x 2 C x C 1/ g5 .x/ D g7 .x/ .x 6 C x 3 C 1/ g7 .x/ D g9 .x/ .x 3 C x 2 C 1/ g9 .x/ D g11 .x/ .x 6 C x 5 C x 3 C x 2 C 1/ g11 .x/ D g13 .x/ .x 6 C x 4 C x 3 C x C 1/ g13 .x/ D g15 .x/ .x 6 C x 5 C x 4 C x 2 C 1/ g15 .x/ D g21 .x/ .x 2 C x C 1/ g21 .x/ D g23 .x/ .x 6 C x 5 C x 4 C x C 1/ g23 .x/ D g27 .x/ .x 3 C x C 1/ g27 .x/ D g31 .x/
k 57 51 45 39 36 30 24 18 16 10 7 d 3 5 7 9 11 13 15 21 23 27 31 t 1 2 3 4 5 6 7 10 11 13 15 Example 11.2.31 Let q D 2, m D 4, and choose as roots , 2 , 3 , 4 , 5 , 6 , with primitive element of H G F.24 /; then a (15,5) code is obtained having dmin 7, and g.x/ D .x 4 C x C 1/.x 4 C 3 C x 2 C x C 1/.x 2 C x C 1/. x
ReedSolomon codes
ReedSolomon codes represent a particular case of BCH codes obtained by choosing m D 1; in other words, the eld G F.q/ and the extension eld G F.q m / coincide. Choosing as a primitive element of (11.180) we get n D qm 1Dq 1 (11.184)
Note that the minimum function with coefcients in G F.q/ of an element i from G F.q/ is m i .x/ D .x i / (11.185)
886
1,
then
1
(11.186)
so that r D .d 1/; the block length n is given by the order of . In this case we show H H that dmin D d; in fact, for any code we have dmin r C 1, as we can always choose a code word with every message symbol but one equal to zero and therefore its weight is at most H equal to r C 1; from this it follows that dmin d, but from the BCH theorem we know H d. that dmin Example 11.2.32 Choose as a primitive element of G F.25 /, and choose the roots , 2 , 3 , 4 , 5 , and H 6 ; then the resultant (31,25) code has dmin D 7, g.x/ D .x /.x 2 /.x 3 /.x 4 /.x 5 /.x 6 /, and the symbols of the code words are from G F.25 /. Observation 11.2 The encoding of ReedSolomon codes can be done by the devices of Figure 11.3 or Figure 11.4, where the operations are in G F.q/. In Table 11.19 and Table 11.20 we give, respectively, the tables of additions and multiplications between elements of G F.q/ for q D
Table 11.19 Addition table for the elements of GF.24 /.
C 0 1
0 0 1
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 0 4 4 8 14 1 10 13 9 2 7 5 12 11 6 3 0 5 9 0 2 11 14 10 3 8 6 13 12 7 0 6 10 1 3 12 0 11 4 9 7 14 13 0 7 11 2 4 13 1 12 5 10 8 0 0 12 8 12 3 5 14 2 13 6 11 9 0 9 9 13 4 6 0 3 14 7 12 0 10 14 5 7 1 4 0 8
0 0
2 2 8 5
3 3 14 9 6
4 4 1 0 10 7 6 6 13 11 3 2
5 5 10 2 1 11 8
7 7 9 14 12 4 3 13 10 0 11 0 6 8 2 5 1 8 8 2 10 0 13 5 4 14 11 0 12 1 7 9 3 6 9 9 7 3 11 1 14 6 5 0 12 0 13 2 8 10 4 10 10 5 8 4 12 2 0 7 6 1 13 0 14 3 9 11 11 11 12 6 9 5 13 3 1 8 7 2 14 0 12 12 11 13 7 10 6 14 4 2 9 8 3 0 0 4 10 0 1 5 0 2 0
13 13 6 12 14 8 11 7 0 5 3 10 9 4 1
14 14 3 7 13 0 9 12 8 1 6 4 11 10 5 2
887
0 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 0 1 0 2 0 3 0 4 0 5 0 6 0 7 0 8 0 9 0 10 0 11 0 12 0 13 0 14 0
0 0 0 0 0
1 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 0 2 0 2 3 4 5 6 7 8 9 10 11 12 13 14 0 1 3 0 3 4 5 6 7 8 9 10 11 12 13 14 0 1 2 4 0 4 5 6 7 8 9 10 11 12 13 14 0 1 2 3 5 0 5 6 7 8 9 10 11 12 13 14 0 1 2 3 4 6 0 6 7 8 9 10 11 12 13 14 0 1 2 3 4 5 7 0 7 8 9 10 11 12 13 14 0 1 2 3 4 5 6 8 0 8 9 10 11 12 13 14 0 1 2 3 4 5 6 7 9 0 9 10 11 12 13 14 0 1 2 3 4 5 6 7 8 10 0 10 11 12 13 14 0 1 2 3 4 5 6 7 8 9 11 0 11 12 13 14 0 1 2 3 4 5 6 7 8 9 10 12 0 12 13 14 0 1 2 3 4 5 6 7 8 9 10 11 13 0 13 14 0 1 2 3 4 5 6 7 8 9 10 11 12 14 0 14 0 1 2 3 4 5 6 7 8 9 10 11 12 13 24 ; the conversion of the symbol representation from binary to exponential is implemented using Table 11.21. We note that the encoding operations can be performed by interpreting the eld elements as polynomials with coefcients from G F.2/, and applying the polynomial arithmetic mod x 4 C x C 1 (see Figure 11.1 and Figure 11.2).
. m 0 /2 . m 0 C1 /2 . m 0 Cd
2 /2
::: :::
1 m 0 Cd
: : : . m 0 Cd
. m 0 /n 1 . m 0 C1 /n 1 : : :
3 7 7 7 5 (11.187)
2 /n 1
888
Table 11.21 Three equivalent representations of the elements of GF.24 /, obtained applying the polynomial arithmetic modulo x4 C x C 1.
Exponential
Polynomial
Binary (x 0 x 1 x 2 x 3 )
0 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14
0 1 x x2 x3 1Cx xC x2 C x2 x3 x3
0000 1000 0100 0010 0001 1100 0110 0011 1101 1010 0101 1110 0111 1111 1011 1001 x3 x3 C
1Cx C xC xC x2
1 C x2 x3 x2 C 1Cx C 1Cx C 1C x2 x3
1 C x2 C x3
we obtain 2 6 sm 0 C1 6 6 Hz D He D s D 6 sm 0 C2 6 : : 4 : sm 0 Cd 2 where sj D
n 1 X `D0
sm 0
3 7 7 7 7 7 5 (11.188)
e` . j /`
j D m 0 ; m 0 C 1; : : : ; m 0 C d
(11.189)
"i .i / j
j D m 0 ; m 0 C 1; : : : ; m 0 C d
(11.190)
889
where the coefcients "i are elements from G F.q m / that represent the values of the errors, and the coefcients i are elements from G F.q m / that give the positions of the errors. In other words, if i D ` then an error has occurred in position `, where ` 2 f0; 1; 2; : : : ; n 1g. The idea of a decoding algorithm is to solve the set of non-linear equations for the unknowns "i and i ; then there are 2 unknowns and d 1 equations. We show that it is possible to solve this set of equations if 2 d 1, assuming m 0 D 1; in the case m 0 6D 1, the decoding procedure does not change. Consider the polynomial in x, also called error indicator polynomial, .x/ D x C
1x 1
C C 1 x C 1
(11.191)
dened as the polynomial that has as zeros the inverse of the elements that locate the positions of the errors, that is i 1 , i D 1; : : : ; . Then .x/ D .1 x1 /.1 x2 / : : : .1 x / (11.192) If the coefcients of .x/ are known, it is possible to nd the zeros of .x/, and thus determine the positions of the errors. The rst step of the decoding procedure consists in evaluating the coefcients 1 ; : : : ; jC using the syndromes (11.190). We multiply both sides of (11.191) by "i i and evaluate 1 the expression found for x D i , obtaining 0 D "i i which can be written as "i .i
jC jC
.1 C 1 i
jC 1
C 2 i
jC 2
C C i /
j
(11.193)
C 1 i
C 2 i
C C i / D 0
(11.194)
(11.194) holds for i D 1; : : : ; , and for every value of j. Adding these equations for i D 1; : : : ; , we get
X i D1
"i .i
jC
C 1 i
jC 1
C 2 i
jC 2
C C i / D 0
for every j
(11.195)
or equivalently
X i D1
"i i
jC
C 1
X i D1
"i i
jC 1
C 2
X i D1
"i i
jC 2
C C
X i D1
As .d 1/=2, if 1 j the equations in (11.196) are equal to the syndromes (11.190); therefore we obtain 1 s jC
1
C 2 s jC
C C s j D
s jC
j D 1; : : : ;
(11.197)
(11.197) is a system of linear equations that can be written in the form 32 3 2 3 2 s3 : : : s 1 s s1 s2 sC1 6 s2 s3 s4 : : : s sC1 7 6 1 7 6 sC2 7 76 7 6 7 6 6 : : : 76 : 7 D 6 : 7 : : 54 : 5 4 : 5 4 : : : : : : s sC1 sC2 : : : s2 2 s2 1 1 s2
(11.198)
890
Thus we have obtained equations in the unknowns 1 ; 2 ; : : : ; . To show that these equations are linearly independent we see that, from (11.190), the matrix of the coefcients can be factorized as 3 2 s1 s2 s3 : : : s 1 s 6 s2 s3 s4 : : : s sC1 7 7 6 6 : : : 7D : : 5 4 : : : : s sC1 sC2 : : : s2 2 s2 1 2 32 32 3 1 1 1 ::: 1 1 1 : : : 1 1 "1 1 6 2 3 : : : 7 6 "2 2 0 7 6 1 2 : : : 2 1 7 6 1 76 76 7 6 : 76 76 : : 7 : 7 :: 6 : : 54 : 5 54 : : : 4 : : :
1 1 2 1 3 1 : : : 1
"
1 : : :
(11.199) The matrix of the coefcients has a non-zero determinant if each of the matrices on the right-hand side of (11.199) has a non-zero determinant. We note that the rst and third matrix are Vandermonde matrices, hence they are non-singular if i 6D m , i 6D m; the second matrix is also non-singular, as it is a diagonal matrix with non-zero terms on the diagonal. This assumes we have at most errors, with .d 1/=2. As is arbitrary, we initially choose D .d 1/=2 and compute the determinant of the matrix (11.199). If this is non-zero, then we have the correct value of ; otherwise, if the determinant is zero, we reduce by one and repeat the computation. We proceed until we obtain a non-zero determinant, thus nding the number of errors that occurred. After nding the solution for 1 ; : : : ; (see (11.198)), we obtain the positions of the errors by nding the zeros of the polynomial .x/ (see (11.192)). Note that an exhaustive method to search for the zeros of the polynomial .x/ requires that at most n possible roots of the type ` ; ` 2 f0; 1; : : : ; n 1g, are taken into consideration. We now compute the value of the non-zero elements "i , i D 1; : : : ; , of the vector e. If the code is binary, the components of the vector e are immediately known, otherwise we solve the system of linear equations (11.190) for the unknowns "1 ; "2 ; : : : ; " . The determinant of the matrix of the system of linear equations is given by 2 6 6 det 6 4 1 2 : : : 2 2 1 2 : : : : : : : : : 1 2 : : : 2 : : : 3 2 7 6 7 6 7 D 1 2 : : : det 6 5 4 1 1 : : : 1 2
1 2 1
::: :::
: : :
1 : : :
1
3 7 7 7 5 (11.200)
The determinant of the Vandermonde matrix in (11.200) is non-zero if errors occurred, as the elements 1 ; : : : ; are non-zero and distinct. In summary, the original system of non-linear equations (11.190) is solved in the following three steps. Step 1: nd the coefcients 1 ; 2 ; : : : ; of the error indicator polynomial by solving a system of linear equations.
891
Step 2: nd the roots 1 ; 2 ; : : : ; of a polynomial of degree . Step 3: nd the values of the errors "1 ; "2 ; : : : ; " by solving a system of linear equations. For binary codes the last step is omitted.
i s j
j D C 1; : : : ; 2
(11.201)
For a given , (11.201) is the equation of a recursive lter, which can be implemented by a shift register with feedback, whose coefcients are given by , as illustrated in Figure 11.8. The solution of (11.198) is thus equivalent to the problem of nding the shift register with feedback of minimum length that, if suitably initialized, yields the sequence of syndromes. This will identify the polynomial .x/ of minimum degree , that we recall exists and is unique, as the matrix of the original problem admits the inverse. The BerlekampMassey algorithm to nd the recursive lter can be applied in any eld and does not make use of the particular properties of the sequence of syndromes s1 ; s2 ; : : : ; sd 1 . To determine the recursive lter we must nd two quantities, that we denote as .L ; .x//, where L is the length of the shift register and .x/ is the polynomial whose degree must satisfy the condition L. The algorithm is inductive, that is for each r, starting from r D 1, we determine a shift register that generates the rst r syndromes.
1 sj Tc sj1
2 Tc
Tc s j
892
The shift register identied by .L r ; .r / .x// will then be a shift register of minimum length that generates the sequence s1 ; : : : ; sr . BerlekampMassey algorithm. Let s1 ; : : : ; sd 1 be a sequence of elements from any eld. Assuming the initial conditions .0/ .x/ D 1, .0/ .x/ D 1, and L 0 D 0, we use the following set of recursive equations to determine .d 1/ .x/: for r D 1; : : : ; d 1, 1r D
n 1 X jD0
.r j
1/
sr
(11.202)
( r D
1 0
if 1r 6D 0 and 2L r otherwise Lr 1
1 / C .1
(11.203) (11.204)
r / L r .r .r
1 1/ .x/
1r 1 r
1r x .1 r / x
(11.205)
1/ .x/
1/ .x/
D1
(11.206)
and sr D
n 1 X jD1
.d j
1/
sr
r D Ld
C 1; : : : ; d
(11.207)
Note that 1r can be zero only if r D 0; in this case we assign to 1r 1 r the value zero. Moreover, we see that the algorithm requires a complexity of the order of d 2 operations, against a complexity of the order of d 3 operations needed by the matrix inversion in (11.198). To prove that the polynomial .d 1/ .x/ given by the algorithm is indeed the .d polynomial of minimum degree with 0 1/ D 1 that satises (11.207), we use the following two lemmas [2]. In Lemma 1 we nd the relation between the lengths of the shift registers of minimum length obtained in two consecutive iterations, L r and L r 1 . In Lemma 2 we use the algorithm to construct a shift register that generates s1 ; : : : ; sr starting from the shift register of minimum length that generates s1 ; : : : ; sr 1 . We will conclude that the construction yields the shift register of minimum length since it satises Lemma 1. Lemma 1. We assume that .L r 1 ; .r 1/ .x// is the shift register of minimum length that generates s1 ; : : : ; sr 1 , while .L r ; .r / .x// is the shift register of minimum length that generates s1 ; : : : ; sr 1 ; sr , and .r / .x/ 6D .r 1/ .x/; then L r max.L r
1; r
Lr
1/
(11.208)
893
Proof. The inequality (11.208) is the combination of the two inequalities L r L r 1 and L r r L r 1 . The rst inequality is obvious, because if a shift register generates a certain sequence it must also generate any initial part of this sequence; the second inequality is obvious if L r 1 r, because L r is a non-negative quantity. Thus we assume L r 1 < r, and suppose that the second inequality is not satised; then L r r 1 L r 1 , or r L r 1 C L r C 1. By assumption we have 8 Lr 1 X .r 1/ > > > sj D i sj i j D L r 1 C 1; : : : ; r 1 > < i D1 (11.209) Lr 1 > X .r 1/ > > > sr 6D i sr i :
i D1
.r / s j k
j D L r C 1; : : : ; r
(11.210)
.r / sr k
Lr X kD1
.r / k
Lr 1 X i D1
.r i
1/
sr
k i
(11.211)
Lr 1 X i D1
1/
sr
1/
.r / sr k
i k
(11.212)
where the expression of sr i is valid as .r i/ goes from r 1 to r L r 1 . Hence it belongs to the set L r C 1; : : : ; r 1, as it is assumed r L r 1 C L r C 1. The summations on the right-hand side of (11.212) can be exchanged, thus obtaining the right-hand side of (11.211). But this yields sr of (11.211) different from sr of (11.212), thus we get a contradiction. Lemma 2. We assume that .L i ; .i / .x//, i D 1; : : : ; r, identies a sequence of shift registers of minimum length such that .i / .x/ generates s1 ; : : : ; si . If .r / .x/ 6D .r 1/ .x/, then L r D max.L r
1; r
Lr
1/
(11.213)
and every shift register that generates s1 ; : : : ; sr , and has a length that satises (11.213), is a shift register of minimum length. The BerlekampMassey algorithm yields this shift register. Proof. From Lemma 1, L r cannot be smaller than the right-hand side of (11.213); thus, if we construct a shift register that yields the given sequence and whose length satises
894
(11.213), then it must be a shift register of minimum length. The proof is obtained by induction. We construct a shift register that satises the Lemma at the r-th iteration, assuming that shift registers were iteratively constructed for each value of the index k, with k r 1. For each k, k D 1; : : : ; r 1, let .L k ; .k/ .x// be the shift register of minimum length that generates s1 ; : : : ; sk . We assume that L k D max.L k
1; k
Lk
1/
k D 1; : : : ; r
(11.214)
Equation (11.214) is veried for k D 0, as L 0 D 0 and L 1 D 1. Let m be the index k at the most recent iteration that required a variation in the length of the shift register. In other words, at the end of the .r 1/-th iteration, m is the integer such that Lr
1
D Lm > Lm ( D
(11.215)
Lr 1 X i D0
.r i 1/ s j i
0 1r
j D Lr j Dr
1; : : : ; r
(11.216)
1/ .x//
.r / .x/ D .r
1/ .x/
(11.217)
If 1r 6D 0, then it is necessary to nd a new shift register. Recall from (11.215) that there was a variation in the length of the shift register for k D m; therefore ( Lm 1 X .m 1/ 0 j D L m 1; : : : ; m 1 i sj i D sj C (11.218) j Dm 1m 6D 0 i D1 and by induction, Lr as L m > L m
1. 1
D L m D max.L m
1; m
Lm
1/
Dm
Lm
(11.219)
.x/
1r 1m 1 x r
1/ .x//
.m
1,
1/
.x/
(11.220)
m .m 1/ .x/]
Lr
and deg[x r Lr
1/
m C Lm
1/
max.L r
1; r
(11.221)
Lr
1/
(11.222)
895
It remains to prove that the shift register .L r ; .r / .x// generates the given sequence. By direct computation we obtain ! Lr X .r / i s j i sj
i D1 Lr 1 X i D1 .r i 1/ s j i
" 1r 1m
1
D sj C ( D 0 1r
sj
r Cm
Lm 1 X i D1
.L i m
1/
# sj
r Cm i
(11.223)
j D L r ; L r C 1; : : : ; r 1r 1m 1 1m D0 j Dr
Therefore the shift register .L r ; .r / .x// generates s1 ; : : : ; sr . In particular, .L d 1 , .d 1/ .x// generates s1 ; : : : ; sd 1 . This completes the proof of Lemma 2. We have seen that the computational complexity for the solution of the system of equations (11.198), that yields the error indicator polynomial, can be reduced by the BerlekampMassey algorithm. We now consider the system of equations (11.190) that yields the values of the errors. The computation of the inverse matrix can be avoided in the solution of (11.190) by applying the Forney algorithm [2]. We recall the expression (11.191) of the error indicator polynomial .x/, that has zeros for x D i 1 , i D 1; : : : ; , given by .x/ D Dene the syndrome polynomial as s.x/ D
d 1 X jD1 Y `D1
.1
x` /
(11.224)
sj x j D
d 1 XX jD1 jD1
"i i x j
(11.225)
and furthermore dene the error evaluator polynomial !.x/ as !.x/ D s.x/ .x/ Proposition. mod x d
1
(11.226)
"i i
Y
j D1 j 6D i
.1
j x/
(11.227)
Proof. From the denition (11.226) of !.x/, we obtain " # #" d 1 XX Y j j !.x/ D "i i x .1 ` x/ mod x d
jD1 i D1 X i D1 `D1
" "i i x .1 i x/
d 1 X jD1
# .i x/
j 1
Y
`D1 ` 6D i
(11.228) .1 ` x/ mod x
d 1
896
id
1 d 1 x /; 1
thus (11.229)
"i i x.1
id
1 d 1
Y
`D1 ` 6D i
.1
` x/
mod x d
1,
is identical to (11.227).
Y
j D1 j 6D i
.1
x j /
(11.230)
!.` 1 / .1 j ` /
1
!.` 1 / ` 1 0 .` 1 /
(11.231)
.1
j ` 1 /
(11.232)
which proves the rst part of equality (11.231). Moreover, from (11.230), we have 0 .` 1 / D `
Y
j D1 j 6D `
.1
j ` 1 /
(11.233)
which proves the second part of equality (11.231). Example 11.2.33 (ReedSolomon (15,9) code with d D 7 (t D 3), and elements from G F.24 /) From (11.186), using Table 11.19 and Table 11.20, the generator polynomial is given by g.x/ D .x /.x 2 /.x 3 /.x 4 /.x 5 /.x 6/ (11.234)
D x 6 C 10 x 5 C 14 x 4 C 4 x 3 C 6 x 2 C 9 x C 6
Suppose that the code polynomial c.x/ D 0 is transmitted, and that the received polynomial is z.x/ D x 7 C 5 x 5 C 11 x 2 (11.235)
897
In this case e.x/ D z.x/. From (11.189), using Table 11.19 and Table 11.20, the syndromes are s1 D 7 C 5 5 C 11 2 D 12 s2 D 14 C 5 10 C 11 4 D 1 s3 D 21 C 5 15 C 11 6 D 14 s4 D 28 C 5 20 C 11 8 D 13 s5 D 35 C 5 25 C 11 10 D 1 s6 D 42 C 5 30 C 11 12 D 11 From (11.202)(11.205), the algorithm develops according to the following d steps, starting from the initial conditions .0/ .x/ D 1 .0/ .x/ D 1 L0 D 0 Step 1 (r D 1): 11 D 12 , and 2L 0 D 0 D r .1/ .x/ 1 D 3 .1/ .x/ 1, 1 D 1, L 1 D 1. 1 1 C 12 x 12 x D 3 1 0 (11.237) 1D6 (11.236)
Step 2 (r D 2): 12 D 1 C 9 D 7 , and 2L 1 D 2 > r 1, 2 D 0, L 2 D L 1 D 1. .2/ .x/ 1 C 12 x 1 7 x D 0 x 3 .2/ .x/ 1 C 12 x C 10 x 1 C 3 x D D 3 x 3 x Step 3 (r D 3): 13 D 14 C 3 D 1, and 2L 2 D 2 D r 1, 3 D 1, L 3 D 3 1 D 2. .3/ .x/ 1 C 3 x C 3 x 2 1 x 1 C 3 x D D 1 0 3 x 1 C 3 x .3/ .x/ Step 4 (r D 4): 14 D 13 C 3 14 C 3 D 1, and 2L 3 D 4 > r 1, 4 D 0, L 4 D L 3 D 2. .4/ .x/ 1 x 1 C 3 x C 3 x 2 D 0 x 1 C 3 x .4/ .x/ 1 C 14 x 1 C 3 x C 3 x 2 C x C 3 x 2 D D x C 3 x 2 x C 3 x 2
898
Step 5 (r D 5): 15 D 1 C 14 13 D 11 , and 2L 4 D 4 D r 1, 5 D 1, L 5 D 5 2 D 3. .5/ .x/ 1 C 14 x 1 C 14 x C 11 x 2 C 14 x 3 1 11 x D D 4 x C 3 x 2 0 4 C 3 x .5/ .x/ Step 6 (r D 6): 16 D 11 C 14 C 11 13 C 14 14 D 0, 6 D 0, L 6 D L 5 D 3. .6/ .x/ 1 0 1 C 14 x C 11 x 2 C 14 x 3 D 0 x 4 C 3 x .6/ .x/ 1 C 14 x C 11 x 2 C 14 x 3 D 4 x C 3 x 2 The error indicator polynomial is .x/ D .6/ .x/. By using the exhaustive method to nd the three roots, we obtain .x/ D 1 C 14 x C 11 x 2 C 14 x 3 D .1 7 x/.1 5 x/.1 2 x/ (11.238)
Consequently the three errors are at positions ` D 2; 5; and 7. To determine the values of the errors we use the Forney algorithm. The derivative of .x/ is given by 0 .x/ D 14 C 14 x 2 The error evaluator polynomial is given by !.x/ D . 12 x C x 2 C 14 x 3 C 13 x 4 C x 5 C 11 x 6 / .1 C 14 x C 11 x 2 C 14 x 3 / mod x 6 D 12 x C 12 x 2 C 8 x 3 Thus the values of the errors are "2 D "5 D "7 D !. 2 / D 11 2 0 . 2 / !. 5 / D 5 5 0 . 5 / !. 7 / D 7 0 . 7 / (11.240) (11.239)
(11.241)
An alternative approach for the encoding and decoding of ReedSolomon codes utilizes the concept of Fourier transform on a Galois eld [2, 5]. Let be a primitive element of the eld G F.q/. The Fourier transform on the eld G F.q/ (GFFT) of a vector c D .c0 ; c1 ; : : : ; cn 1 / of n bits is dened as .C0 ; C1 ; : : : ; C n 1 /, where Cj D
n 1 X i D0
ci i j
j D 0; : : : ; n
(11.242)
899
Let us consider a code word c of n bits in the time domain from a ReedSolomon cyclic code that corrects up to t errors; then c corresponds to a code polynomial that has as roots 2t D d 1 consecutive powers of . If we take the GFFT of this word, we nd that in the frequency domain the transform has 2t consecutive components equal to zero. Indeed from (11.176), specialized to ReedSolomon codes, and from (11.242), we can show that the two conditions are equivalent, that is a polynomial has 2t consecutive powers of as roots if and only if the transform has 2t consecutive components equal to zero. The approach that resorts to the GFFT is therefore the mirror of the approach that uses the generator polynomial. This observation leads to the development of efcient methods for encoding and decoding.
11.2.6
In this section we consider the probability of error in the decoding of block codes, in the case of decoding with hard or soft input (see Section 6.8). For an in-depth study of the subject we refer the reader, for example, to [6]. With reference to Figure 6.20, let Pbit be the bit error probability for the detection of the bits of the binary sequence fcm g, or bit error probability of the channel, Pw the error Q .dec/ probability for a code word, and Pbit the error probability for a bit of the binary sequence H O fbl g obtained after decoding. For a .n; k/ block code with t D .dmin 1/=2 and hard input decoding the following inequality holds: n X n i Pbit .1 Pbit /n i Pw (11.243) i i DtC1 which, under the condition n Pbit 1, can be approximated as n P tC1 .1 Pbit /n t 1 Pw ' t C 1 bit
(11.244)
The inequality (11.243) follows from the channel model (11.12) assuming errors that are i.i.d., and from the consideration that the code may not be perfect (see page 839), and therefore it could correct also some received words with more than t errors. If a word error occurs, the most probable event is that the decoder decides for a code H H word with distance dmin D 2t C 1 from the transmitted code word, thus making dmin bit errors in the sequence fcm g. As c is formed of n bits, we have that at the decoder output O the bit error probability is
.dec/ Pbit '
2t C 1 Pw n
(11.245)
Example 11.2.34 H For a (5,1) repetition code with dmin D 5 (see page 839), decoding with hard input yields 5 5 3 5 Pw D Pbit .1 Pbit /2 C P 4 .1 Pbit / C Pbit (11.246) 3 4 bit
900
Example 11.2.35 H For an .n; k/ Hamming code with dmin D 3 (see page 839), (11.243) yields Pw
n X n i D2 i Pbit .1 n
i [.1
Pbit /n
(11.247) Pbit /
n 1
D1
Pbit / C n Pbit .1
For example, for a (15,11) code, if Pbit D 10 3 then Pw ' 10 4 , and from (11.245) we .dec/ get Pbit ' 2 10 5 . The decoders that have been considered so far are classied as hard input decoders, as the demodulator output is quantized to the values of the coded symbols before decoding. In general, other decoding algorithms with soft input may be considered, that directly process the demodulated signal, and consequently the decoder input is real valued (see Section 11.3.2). In the case of antipodal binary signals and soft input decoding we obtain (see also Section 6.8 on page 496) 0s 1 H Rc dmin 2E b A (11.248) Pw ' .2k 1/ Q @ N0
11.3
Convolutional codes
Convolutional codes are a subclass of the class of tree codes, so called because their code words are conveniently represented as sequences of nodes in a tree. Tree codes are of great interest because decoding algorithms have been found that are easy to implement, and can be applied to the entire class of tree codes, in contrast to decoding algorithms for block codes, each designed for a specic class of codes, as for example BCH codes. Several approaches have been used in the literature for describing convolutional codes; here we will illustrate these approaches by rst considering a specic example. Example 11.3.1 Consider a rate 1=2 binary convolutional code, obtained by the encoder illustrated in .1/ .2/ Figure 11.9a. For each bit bk that enters the encoder, two output bits, ck and ck , .1/ are transmitted. The rst output ck is obtained if the switch at the output is in the .2/ upper position, and the second output ck is obtained if the switch is in the lower position; the two previous input bits, bk 1 and bk 2 , are stored in the memory of the encoder. As the information bit is not presented directly to one of the outputs, we say that the code is nonsystematic. The two coded bits are generated as linear combinations of the bits of the message; denoting the input sequence as : : : ; b0 ; b1 ; b2 ; b3 ; : : : , and .1/ .2/ .1/ .2/ .1/ .2/ .1/ .2/ the output sequence as : : : ; c0 ; c0 ; c1 ; c1 ; c2 ; c2 ; c3 ; c3 ; : : : , then the following
901
ck(1) bk
ck
(a)
(2)
10 10 01 (a) 01 11 (c) 00 10 1 11 (d) 01 0 11 (c) 10 00 (d) 11 00 (d) 00 (d) 00 (c) 10 11 (b) 11 01 (a) 01 00 (d) 00 10 (b) (c) 10 11 (b) 11 01 (a) 01 00
(a)
(b) (c)
(d) (a)
(b) (c)
(d) (a)
(b) (c)
(d) (a)
(b) (c)
(d)
(b)
Figure 11.9. (a) Encoder and (b) tree diagram for the convolutional code of Example 11.3.1.
902
relations hold:
.1/ c k D bk bk .2/ ck 1 2
bk
(11.249)
D bk bk
A convolutional code may be described in terms of a tree, trellis, or state diagram; for the code dened by (11.249) these descriptions are illustrated in Figures 11.9b, 11.10a, and 11.10b, respectively. With reference to the tree diagram of Figure 11.9b, we begin at the left (root) node and proceed to the right by choosing an upper path if the input bit is equal to 1 and a lower path if the input bit is 0. We output the two bits represented by the label on the branch that takes us to the next node, and then repeat this process at the next node. The nodes or
(a)
(b)
Figure 11.10. (a) Trellis diagram and (b) state diagram for the convolutional code of Example 11.3.1.
903
states of the encoder are labeled with the letters a, b, c, and d, which indicate the relation with the four possible values assumed by the two bits stored in the encoder, according to the table: bk 0 1 0 1
1
bk 0 0 1 1
label d c b a (11.250)
If for example we input the sequence b0 ; b1 ; b2 ; b3 ; D 1 1 0 1 : : : , we would then .1/ .2/ .1/ .2/ .1/ .2/ .1/ .2/ output the sequence c0 ; c0 ; c1 ; c1 ; c2 ; c2 ; c3 ; c3 ; D 1 1 0 1 0 1 0 0 : : : . As, at any depth in the tree, nodes with the same label will have the same tree growing from them, we can superimpose these nodes on a single node. This results in the trellis diagram represented in Figure 11.10a, where solid and dashed lines correspond to transitions determined by input bits equal to 1 and 0, respectively. The state diagram for the encoder is illustrated in Figure 11.10b. The four states (a, b, c, d) correspond to the four possible combinations of bits stored in the encoder. If the encoder is in a certain state, a transition to one of two possible states occurs, depending on the value of the input bit. Possible transitions between states are represented as arcs, on which an arrow indicates the direction of the transition; with each arc is associated a label that indicates the value assumed by the input bit, and also the value of the resulting output bits. The description of the encoder by the state diagram is convenient for analyzing the properties of the code, as we will see later. It is also convenient to represent code sequences in terms of the D transform, as b.D/ D b0 C b1 D C b2 D 2 C b3 D 3 C
.1/ .1/ .1/ .1/ c.1/ .D/ D c0 C c1 D C c2 D 2 C c3 D 3 C D g .1;1/ .D/ b.D/ .2/ .2/ .2/ .2/ c.2/ .D/ D c0 C c1 D C c2 D 2 C c3 D 3 C D g .2;1/ .D/ b.D/
(11.251)
11.3.1
In general we consider convolutional codes with symbols from G F.q/; assuming the encoder produces n 0 output code symbols for every k0 input message symbols, the code rate is equal to k0 =n 0 . It is convenient to think of the message sequence as being the interlaced version of k0 different message sequences, and to think of the code sequence as the interlaced version of n 0 different code sequences. In other words, given the information sequence fb` g we form the k0 subsequences
.i bk / D bkk0 Ci 1
i D 1; : : : ; k0
(11.252)
904
(11.253)
Let c.1/ .D/; c.2/ .D/; : : : ; c.n 0 / .D/ be the D transforms of the n 0 output sequences; then c. j/ .D/ D
k0 X i D1
j D 1; 2; : : : ; n 0
(11.254)
An .n 0 ; k0 / convolutional code is then specied by giving the coefcients of all the polynomials g . j;i / .D/, i D 1; 2; : : : ; k0 , j D 1; 2; : : : ; n 0 . If for all j D 1; 2; : : : ; k0 , we have ( 1 j Di (11.255) g . j;i / .D/ D 0 j 6D i then the code is systematic and k0 of the n 0 output sequences are just the message sequences. An encoder for a convolutional code needs storage elements. Let be the constraint length of the code,5 D max.deg g . j;i / .D//
j;i
(11.256)
Therefore the encoder of a convolutional code must store previous blocks of k0 message symbols to form a block of n 0 output symbols. The general structure of an encoder for a code with k0 D 1 and n 0 D 2 is illustrated in Figure 11.11; for such an encoder, k0 storage elements are necessary. If the code is systematic, then the encoder can be implemented with .n 0 k0 / storage elements, as illustrated in Figure 11.12 for k0 D 2 and n 0 D 3. If we interpret the sequence fck g as the output of a sequential nite-state machine (see Appendix 8.D), at instant k the trellis of a nonsystematic code is dened by the three signals: 1. Input 2. State 3. Output
. j/ .1/ .2/ .k [bk ; bk ; : : : ; bk 0 / ] .k / .k / .1/ .1/ [bk ; : : : ; bk 0 ; : : : ; bk . 1/ ; : : : ; bk 0. 1/ ] .1/ .2/ .n [ck ; ck ; : : : ; ck 0 / ]
where ck ; j D 1; : : : ; n 0 , is given by (11.254). Then there are q k0 states in the trellis. There are q k0 branches departing from each state and q k0 branches merging into a state. The output vector consists of n 0 qary symbols.
5
905
c (1)(D)
g0
(1,1)
g1
(1,1)
g2
(1,1)
(1,1) 3
(1,1)
(1)
(D)
g (2,1)
0
D
g (2,1)
1
D
g (2,1)
2
D
g (2,1)
3
D
g (2,1)
c (2)(D)
Figure 11.11. Block diagram of an encoder for a convolutional code with k0 D 1, n0 D 2, and constraint length .
(1) b (D)
b (D)
(2)
g (3,1)
1
g 1
(3,2)
(3,1) 0
(3,2)
0
c (3)(D)
Figure 11.12. Block diagram of an encoder for a systematic convolutional code with k0 D 2, n0 D 3, and constraint length .
906
::: :::
.n 0 ;1/ .n 0 ;2/
(11.260)
g0
.1;k0 / .2;k0 / .n ;k / g0 : : : g0 0 0
g1
: : : g1
.n 0 ;k0 /
: : : g
.1;k0 / .2;k0 / .n ;k / g : : : g 0 0
If the code is systematic, the parity matrix of the generator polynomials can be written as I P0 0 P1 : : : 0 P where I and 0 are k0 k0 matrices and Pi , i The semi-innite parity check matrix is then 2 T 0 P0 I 6 PT 0 T P0 6 6 : : :1 ::: 6 H1 D 6 T T 6 P 0 P 1 6 4 0 0 T P ::: ::: D 0; : : : ; , are k0 .n 0 0 I 0 0 0 0 :::
T P T P ::: 2 1
(11.261) k0 / matrices.
(11.262)
Thus for any code word c of innite length, H1 c D 0. Often, rather than considering the semi-innite matrix H1 , we consider the nite matrix H dened as 2 3 T P0 I 0 0 ::: 0 0 6 7 T T P0 I : : : 0 0 7 6 P1 0 (11.263) HD6 7 ::: ::: ::: 4 ::: 5 T T T P 0 P 1 0 : : : P0 I The bottom row of matrices of the matrix H is called the basic parity check matrix. From it we can see that the parity symbols in a block are given by the linear combination of T information bits in that block, corresponding to non-zero terms in P0 , in the immediately T , and so on until the -th preceding preceding block, corresponding to non-zero terms in P1 T block, corresponding to non-zero terms in P .
Generator matrix
From (11.260), we introduce the matrices 2 .1;1/ .n ;1/ 3 gi gi.2;1/ : : : gi 0 6 : 7 : : gi D 4 : 5 : : .1;k0 / .2;k0 / .n 0 ;k0 / gi : : : gi gi
i D 0; : : : ;
(11.264)
Hence the generator matrix is of the form 2 3 g0 g1 : : : g 0 ::: G1 D 4 0 g0 : : : g 1 g : : : 5 ::: ::: ::: ::: ::: :::
(11.265)
907
Some examples of convolutional codes with the corresponding encoders and generator matrices are illustrated in Figure 11.13.
Transfer function
H An important parameter of a convolutional code is dfree , that determines the performance of the code (see Section 11.3.3).
Denition 11.14 Let e.D/ D [e.n 0 / .D/; : : : ; e.1/ .D/], be any error sequence between two code words .n .1/ .n .1/ c1 .D/ D [c1 0 / .D/; : : : ; c1 .D/] and c2 .D/ D [c2 0 / .D/; : : : ; c2 .D/], that is c1 .D/ D .n .1/ c2 .D/ C e.D/, and ek D [ek 0 / ; : : : ; ek ] denotes the k-th element of the sequence. We dene the free Hamming distance of the code as
H dfree D min e.D/ 1 X kD0
w.ek /
(11.266)
H where w is introduced in Denition 11.4 on page 832. As the code is linear, dfree corresponds to the minimum number of symbols different from zero in a non-zero code word.
Next we consider a method to compute the weights of all code words in a convolutional code; to illustrate the method we examine the simple binary encoder of Figure 11.9a. We begin by reproducing the trellis diagram of the code in Figure 11.14, where each path is now labeled with the weight of the output bits corresponding to that path. We consider all paths that diverge from state (d) and return to state (d) for the rst time after a number of steps j. By inspection, we nd one such path of weight 5 returns to state (d) after 3 steps; moreover, we nd two distinct paths of weight 6, one that returns to state (d) after 4 steps and another after 5 steps. Hence we nd that this code has H dfree D 5. We now look for a method that enables us to nd the weights of all code words as well as the lengths of the paths that give origin to the code words with these weights. Consider the state diagram for this code, redrawn in Figure 11.15 with branches labeled as D 2 , D, or D 0 D 1, where the exponent corresponds to the weight of the output bits corresponding to that branch. Next we split node (0,0) to obtain the state diagram of Figure 11.16, and we compute a generating function for the weights. The generating function is the transfer function of a signal ow graph with unit input. From Figure 11.16, we obtain this transfer function by solving the system of equations D D2 C 1 D D C D D D C D D D2 (11.267)
908
g1 =(1,0) g =(1,1)
2
k 0 =1, n0 =2, =2
(a)
c (2) (D)
(1)
(D)
D c
(1)
(2)
(D)
c b
(2)
(3)
(D)
(D)
D k 0 =2, n0 =3, =1
(b)
c b
(1)
(1)
(D)
(D)
D c
(2)
(D)
b (2) (D)
g0 =
D c (3) (D)
b (3) (D)
g1 =
c (4) (D)
k0 =3, n 0 =4, =1
(c)
909
Figure 11.14. Trellis diagram of the code of Example 11.3.1; the labels represent the Hamming weight of the output bits.
Figure 11.15. State diagram of the code of Example 11.3.1; the labels represent the Hamming weight of the generated bits.
Figure 11.16. State diagram of the code of Example 11.3.1; node (0,0) is split to compute the transfer function of the code.
910
Figure 11.17. State diagram of the code of Example 11.3.1; node (0,0) is split to compute the augmented transfer function.
From inspection of t .D/, we nd there is one code word of weight 5, two of weight 6, four of weight 7, : : : . Equation (11.268) holds for code words of innite length. If we want to nd code words that return to state (d) after j steps we refer to the state diagram of Figure 11.17. The term L introduced in the label on each branch allows to keep track of the length of the sequence, as the power of L is augmented by 1 every time a transition occurs. Furthermore, we introduce the term I in the label on a branch if the corresponding transition is due to an information bit equal to 1; this allows computation for each path on the trellis diagram of the corresponding number of information bits equal to 1. The augmented transfer function is given by t .D; L ; I / D D5 L 3 I D L.1 C L/I (11.269)
Thus we see that the code word of weight 5 is of length 3 and is originated by a sequence of information bits that contains one bit equal to 1, there are two code words of weight 6, one of length 4 and the other of length 5, both of which are originated by a sequence of information bits that contain two bits equal to 1, : : : .
911
Figure 11.18. (a) Encoder and (b) state diagram for a catastrophic convolutional code.
state .1; 1/ does not increase the weight of the code word, so that a code word corresponding to a path passing through the states .0; 0/; .1; 0/; .1; 1/; .1; 1/; : : : ; .1; 1/; .0; 1/; .0; 0/ is of weight 6, independently of the number of times it passes through the self loop at state (1,1). In other words, long sequences of coded bits equal to zero may be obtained by remaining in the state .0; 0/ with a sequence of information bits equal to zero, or by remaining in the state .1; 1/ with a sequence of information bits equal to one. Therefore a limited number of channel errors, in this case 6, can cause a large number of errors in the sequence of decoded bits. Denition 11.15 A convolutional code is catastrophic if there exists a closed loop in the state diagram that has all branches with zero weight.
912
~ (1,1) (D) g
~ ( n0 ,1)(D) g
(a)
g b (1)(D)= 1 gc (D) g
(1,1)
(D)
~ (1,1) (D) g
( n0 ,1)
(D)
~ ( n0 ,1)(D) g
(b)
Figure 11.19. Two distinct innite sequences of information bits that produce the same output sequence with a nite number of errors.
For codes with rate 1=n 0 , it has been shown that a code is catastrophic if and only if all generator polynomials have a common polynomial factor. In the above example, the common factor is 1 C D. This can be proved using the following argument: suppose that g .1;1/ .D/; g .2;1/ .D/; : : : ; g .k0 ;1/ .D/ all have the common factor gc .D/, so that Q g .i;1/ .D/ D gc .D/ g .i;1/ .D/ (11.270)
Suppose the all zero sequence is sent, b.1/ .D/ D 0, and that the nite error sequence g .i;1/ .D/, equal to that dened in (11.270), occurs in the i-th subsequence output, for Q i D 1; 2; : : : ; n 0 , as illustrated in Figure 11.19a. The same output sequence is obtained if the sequence of information bits with innite length b.1/ .D/ D 1=gc .D/ is sent, and no channel errors occur, as illustrated in Figure 11.19b. Thus a nite number of errors yields a decoded sequence of information bits that differ from the transmitted sequence in an innite number of positions.
11.3.2
Various algorithms have been developed for the decoding of convolutional codes. One of the rst decoding methods was algebraic decoding, which is similar to the methods developed for the decoding of block codes. However, this method has the disadvantages that it is applicable only to a limited number of codes having particular characteristics, and exhibits
913
performance that is lower as compared to decoding methods based on the observation of the whole received sequence. The latter methods, also called probabilistic decoding methods, include the Viterbi algorithm (VA), the sequential decoding algorithm by Fano [6], and the forward-backward algorithm by BahlCockeJelinekRaviv (BCJR). Before illustrating the various decoding methods, we consider an important function.
Interleaving
The majority of block codes as well as convolutional codes are designed by assuming that the errors introduced by the noisy channel are statistically independent. This assumption is not always true in practice. To make the channel errors, at least approximately, statistically independent it is customary to resort to an interleaver, which performs a permutation of the bits of a sequence. For example, a block interleaver orders the coded bits in a matrix with M1 rows and M2 columns. The coded bits are usually written in the matrix by row and then read by column before being forwarded to the bit mapper. At the receiver, a deinterleaver stores the detected bits in a matrix of the same M1 M2 dimensions, where the writing is done by column and the reading by row. As a result, possible error bursts of length M1 B are broken up into bursts of shorter length B.
(11.271)
At the receiver, two models may be adopted. Model with hard input. With reference to the transmission system of Figure 6.20, we consider the sequence at the output of the binary channel. In this case the demodulator has already detected the transmitted symbols, for example, by a threshold detector, and the inverse bit mapper provides the binary sequence fz m D cm g to the decoder, from which we Q obtain the interlaced binary sequences
.1/ .1/ .1/ z k D z 2k D ck ek .2/ .2/ .2/ z k D z 2kC1 D ck ek
(11.272)
.i where the errors ek / 2 f0; 1g, for a memoryless binary symmetric channel, are i.i.d. (see (6.91)). From the description on page 904, introducing the state of the encoder at instant k as the vector with elements
sk D [bk ; : : : ; bk
. 1/ ]
(11.273)
914
the desired sequence in (11.272), that coincides with the encoder output, can be written as (see (11.271) and (11.273))
.1/ ck D f .1/ .sk ; sk .2/ ck D f .2/ .sk ; sk 1/ 1/
(11.274)
Model with soft input. Again with reference to Figure 6.20, at the decision point of the receiver the signal can be written as (see (8.173)) z k D u k C wk (11.275)
2 where we assume wk is white Gaussian noise with variance w D 2 I2 , and u k is given by (8.174) L2 X nD L 1
uk D
n ak
(11.276)
where fak g is the sequence of symbols at the output of the bit mapper. Note that in (11.276) the symbols fak g are in general not independent, as the input of the bit mapper is a code sequence according to the law (11.274). The relation between u k and the bits fb` g depends on the intersymbol interference in (11.276), the type of bit mapper and the encoder (11.271). We consider the case of absence of ISI, that is u k D ak (11.277)
.1/ .2/ and a 16-PAM system where, without interleaving, four consecutive code bits c2k , c2k , .1/ .2/ c2k 1 , c2k 1 are mapped into a symbol of the constellation. For an encoder with constraint length we have .1/ .2/ .1/ .2/ u k D fQ.ak / D fQ[BMAP f[c2k ; c2k ; c2k 1 ; c2k 1 ]g]
/; g
.2;1/
.b2k ; : : : ; b2k
(11.278)
1 ; : : : ; b2k 1 /; g
.2;1/
.b2k
1 ; : : : ; b2k 1 /g]
C1 ]
(11.279)
1/
(11.280)
We observe that in this example each state of the trellis admits four possible transitions. As we will see in Chapter 12, better performance is obtained by jointly optimizing the encoder and the bit mapper.
915
Viterbi algorithm
The Viterbi algorithm, described in Section 8.10.1, is a probabilistic decoding method that implements the maximum likelihood criterion, which minimizes the probability of detecting a sequence that is different from the transmitted sequence. VA with hard input. The trellis diagram is obtained by using the denition (11.273), and .1/ .2/ .1/ .2/ the branch metric is the Hamming distance between zk D [z k ; z k ]T and ck D [ck ; ck ]T , (see Denition 11.1), d H .zk ; ck / D number of positions where zk differs from ck where ck is generated according to the rule (11.274). VA with soft input. The trellis diagram is now obtained by using the denition (11.279), and the branch metric is the Euclidean distance between z k and u k , jz k u k j2 (11.282) (11.281)
where u k , in the case of the previous example of absence of ISI and 16-PAM transmission, is given by (11.280). As an alternative to the VA we can use the FBA of Section 8.10.2.
Forward-backward algorithm
The previous approach, which considers joint detection in the presence of ISI and convolutional decoding, requires a computational complexity that in many applications may turn out to be exceedingly large. In fact, the state (11.279), that takes into account both encoding and the presence of ISI, usually is difcult to dene and is composed of several bits of the sequence fb` g. An approximate solution is obtained by considering the detection and the decoding problems separately, however, assuming that the detector passes the soft information on the detected bits to the decoder. Soft output detection by FBA. By using a trellis diagram that takes into account the ISI introduced by the channel, the code bits fcn g are detected assuming that they are i.i.d., and the reliability of the detection is computed (soft detection). For this purpose we use the FBA of page 670, that determines for each state a metric Vk .i/, i D 1; : : : ; Ns . Now, with reference to the example of the channel given by (11.277) and 16-PAM transmission, the state is identied by sk D .ak / D [c4k ; c4k 1 ; c4k 2 ; c4k 3 ], where6 fcn g, cn 2 f 1; 1g, is assumed to be a sequence of i.i.d. symbols. By considering the binary state representation, and by suitably adding the values Vk .i/, we get the MAP metric, or
It is sometimes convenient to view the encoder output cn and/or the encoder input bn as symbols from the alphabet f 1; C1g, rather than f0; 1g. It will be clear from the context to which alphabet we refer.
916
Vk .i/
2 f 1; 1g
t D 0; 1; 2; 3
(11.283)
2 f 1; 1g
(11.284)
By the above formulation, the soft decision associated with the bit cn is given by
.in/ .in/ `n D `n .1/ .in/ `n . 1/
(11.285)
also called log-likelihood ratio (LLR). Observation 11.3 For binary transmission in the absence of ISI, from (8.269) on page 675, we have, apart from a non-essential additive constant,
.in/ `n ./ D
.z n /2 2 I2
2 f 1; 1g
(11.286)
2 zn I2
(11.287)
In other words, apart from a constant factor, the LLR associated with the bit cn coincides with the demodulator output z n . Rather than (11.284) and (11.283), we can use the Max-Log-MAP criterion (8.267) that yields an approximate log-likelihood, Q `.in/ t ./ D 4k
i 2 f1; : : : ; Ns g i with t-th binary component equal to
max
vk .i/
(11.288)
An alternative to the FBA is obtained by modifying the VA to yield a soft output (SOVA), as discussed in the next section. Convolutional decoding with soft input (SI). The decoder for the convolutional code typically uses the VA with branch metric (associated with a cost function to be minimized) given by jj
.in/ k
ck jj
(11.289)
917
.in/ k
.1/ .2/ LLR associated, respectively, with ck and ck . As jck j D 1, (11.289) can be rewritten as:
.`.in;1/ k
.1/ ck / C .`.in;2/ k 2
.2/ ck /
D .`.in;1/ / C .`.in;2/ / C 2 k k
(11.290)
Leaving out the terms that do not depend on ck , and extending the formulation to a convolutional code with rate k0 =n 0 , the branch metric (associated with a cost function to be maximized) is expressed as (see also [7]) 2 where the factor 2 can be omitted. Observation 11.4 As we have previously stated, best system performance is obtained by jointly designing the encoder and the bit mapper. However in some systems, typically radio, an interleaver is used between the encoder and the bit mapper. In this case joint detection and decoding are impossible to implement in practice. Detection with soft output followed by decoding with soft input remains a valid approach, obviously after re-ordering the LLR as determined by the deinterleaver. In applications that require a soft output (see Section 11.6), the decoder, that is called in this case soft-input soft-output (SISO), can use one of the versions of the FBA or the SOVA.7
n0 X jD1
ck ` k
. j/ .in; j/
(11.291)
Sequential decoding
Sequential decoding of convolutional codes represented the rst practical algorithm for ML decoding. It has been employed, for example, for the decoding of signals transmitted by deep-space probes, such as the Pioneer, 1968 [10]. There exist several variants of sequential decoding algorithms, that are characterized by the search of the optimum path in a tree diagram (see Figure 11.9b), instead of along a trellis diagram, as considered, e.g., by the VA. Sequential decoding is an attractive technique for the decoding of convolutional codes and trellis codes if the number of states of the encoder is large [11]. In fact, as the implementation complexity of ML decoders such as the Viterbi decoder grows exponentially with the constraint length of the code, , the complexity of sequential decoding algorithms is essentially independent of . On the other hand, sequential decoding presents the drawback that the number of computations Nop required for the decoding process to advance by one branch in the decoder tree is a random variable with a Pareto distribution, i.e. P[Nop > N ] D AN
7
(11.292)
An extension of SISO decoders for the decoding of block codes is found in [8, 9].
918
where A and are constants that depend on the channel characteristics and on the specic code and the specic version of sequential decoding used. Real-time applications of sequential decoders require buffering of the received samples. As practical sequential decoders can perform only a nite number of operations in a given time interval, resynchronization of the decoder must take place if the maximum number of operations that is allowed for decoding without incurring buffer saturation is exceeded. To determine whether it is practical for a receiver to adopt sequential decoding, we recall the denition of cut-off rate of a transmission channel, and the associated minimum signalto-noise ratio .E b =N0 /0 (see page 509). Sequential decoders exhibit very good performance, with a reduced complexity as compared to the VA, if the constraint length of the code is sufciently large and the signal-to-noise ratio is larger than .E b =N0 /0 . If the latter condition is not veried, the average number of operations required to produce one symbol at the decoder output is very large. The Fano Algorithm In this section we consider sequential decoding of trellis codes, a class of codes that will be studied in detail in Chapter 12. However, the algorithm can be readily .1/ .k extended to convolutional codes. At instant k, the k0 information bits bk D [bk ; : : : ; bk 0 / ] are input to a rate k0 = .k0 C 1/ convolutional encoder with constraint length that outputs .0/ .k the coded bits ck D [ck ; : : : ; ck 0 / ]. The k0 C 1 coded bits select from a constellation with k0 C1 elements a symbol a , which is transmitted over an additive white Gaussian M D2 k noise channel. Note that the encoder tree diagram has 2k0 branches that correspond to the values of bk stemming from each node. Each branch is labeled by the symbol ak selected by the vector ck . The received signal is given by (see (11.275)) z k D a k C wk (11.293)
The received signal sequence is input to a sequential decoder. Using the notation of Section 8.10.1, in the absence of ISI and assuming a D , the ML metric to be maximized can be written as [6] 2 3 0./ D
K 16 X kD0
7 B7 5
(11.294)
where B is a suitable constant that determines a tradeoff between computational complexity and performance and is related to the denominator in (11.294). Choosing B D k0 1 and Pak .i / D M D 2 .k0 C1/ , i 2 A, we obtain 2 0./ D
K 16 X6 kD0
3
jk k j2 2 I2 jk i j2 2 I2
e 6log 6 2 6 X 4 e
i 2A
7 7 7 C 17 7 7 5
(11.295)
919
Various algorithms have been proposed for sequential decoding [12, 13, 14]. We will restrict our attention here to the Fano algorithm [6, 11]. The Fano algorithm examines only one path of the decoder tree at any time using the metric in (11.294). The considered path extends to a certain node in the tree and corresponds to a segment of the entire code sequence , up to symbol k . Three types of moves between consecutive nodes on the decoder tree are allowed: forward, lateral, and backward. On a forward move, the decoder goes one branch to the right in the decoder tree from the previously hypothesized node. This corresponds to the insertion of a new symbol kC1 in (11.294). On a lateral move, the decoder goes from a path on the tree to another path differing only in the last branch. This corresponds to the selection of a different symbol k in (11.294). The ordering among the nodes is arbitrary, and a lateral move takes place to the next node in order after the current one. A backward move is a move one branch to the left on the tree. This corresponds to the removal of the symbol k from (11.294). To determine which move needs to be made after reaching a certain node, it is necessary to compute the metric 0k of the current node being hypothesized, and consider the value of the metric 0k 1 of the node one branch to the left of the current node, as well as the current value of a threshold T h, which can assume values that are multiples of a given constant 1. The transition diagram describing the Fano algorithm is illustrated in Figure 11.20. Typically, 1 assumes values that are of the order of the minimum distance between symbols. As already mentioned, real-time applications of sequential decoding require buffering of the input samples with a buffer of size S. Furthermore, the depth of backward search is also nite and is usually chosen to be at least ve times the constraint length of the code. To avoid erasures of output symbols in case of buffer saturation, in [15] a buffer looking algorithm (BLA) is proposed. The buffer is divided into L sections, each with size S j ; j D 1; : : : ; L. A conventional sequential decoder (primary decoder) and L 1 secondary decoders are used. The secondary decoders employ fast algorithms, such as the M-algorithm [16], or variations of the Fano algorithm that are obtained by changing the value of the bias B in the metric (11.294). Example 11.3.2 (Sequential decoding of a 512-state 16-PAM trellis code) We illustrate sequential decoding with reference to a 512-state 16-PAM trellis code specied for SHDSL transmission (see Chapter 17). The encoder for this trellis code comprises a rate 1=2 nonsystematic non-recursive convolutional encoder with constraint length D 9 and a bit mapper as specied in Figure 11.21. The symbol error probabilities versus signal-to-noise ratio 0 obtained by sequential decoding with innite buffer size and depth of backward search of 64 and 128 symbols, and by a 512state VA decoding with length of the path memory of 64 and 128 symbols are shown in Figure 11.22. Also shown for comparison are the error probabilities obtained for uncoded 8-PAM and 16-PAM transmission.
11.3.3
H For binary convolutional codes with free distance dfree , and bit error probability of the channel equal to Pbit , decoding with hard input yields .dec/ Pbit ' A 2dfree Pbit
H
(11.296)
920
and decoding with soft input, for a system with antipodal signals, yields 0s
.dec/ Pbit H Rc dfree 2E b
1 A (11.297)
' AQ@
N0
where A is a constant [17]. In particular, we consider BPSK transmission over an ideal AWGN channel. Assuming an encoder with rate Rc D 1=2 and constraint length D 6, the coding gain for a soft Viterbi decoder is about 3.5 dB for Pbit D 10 3 ; it becomes about 4.6 dB for Pbit D 10 5 . Note that a soft decoder allows gain of about 2.4 dB with respect to a hard decoder, for Pbit < 10 3 .
921
Figure 11.21. (a) Block diagram of the encoder and bit mapper for a trellis code for 16-PAM transmission, (b) structure of the rate-1/2 convolutional encoder, and (c) bit mapping for the 16-PAM format.
11.4
Concatenated codes
Concatenated coding is usually introduced to achieve an improved error correction capability [18]. Interleaving is also commonly used in concatenated coding schemes, as illustrated in Figure 11.23, so that the decoding processes of the two codes (inner and outer) can be considered approximately independent. The rst decoding stage is generally utilized to produce soft decisions on the information bits, that are passed to the second decoding stage. While the forward-backward algorithm directly provides a soft output (see (11.285)), the Viterbi algorithm must be slightly modied.
922
10
10
16PAM, uncoded
10
10 P
e
8PAM, uncoded
4
10
10
10
10
SD, depth=64 SD, depth=128 VA, path mem.=64 VA, path mem.=128 18 20 22 (dB) 24 26 28
16
Figure 11.22. Symbol error probabilities for the 512-state 16-PAM trellis code with sequential decoding (depth of search limited to 64 or 128) and 512-state Viterbi decoding (length of path memory limited to 64 or 128). Symbol error probabilities for uncoded 8-PAM and 16-PAM transmission are also shown.
In this section we consider different methods to generate the soft output. The difference metric algorithm (DMA). Figure 11.24 shows a section of a trellis diagram with four states, where we assume sk D .bk ; bk 1 /. We consider two states at instant k 1 that differ in the least signicant bit bk 2 of the binary representation, that is s.0/1 D .00/ k
923
and s.1/1 D .01/. A transition from each of these two states to state sk D .00/ at instant k is k allowed. According to the VA, we choose as survivor sequence the sequence that minimizes the metric .i/ sk 1 !sk .i / (11.298) min 0k 1 .sk 1 / C k
i 2f0;1g
where 0k
.i / 1 .sk 1 /
is the path metric associated with the survivor sequence up to state s.i / 1 k
s
.i/
at instant k 1, and k k 1 denotes the branch metric associated with the transition from state s.i / 1 to state sk . Let Ak D 0k 1 .00/ C k00!00 and Bk D 0k 1 .01/ C k01!00 , then k we choose the upper or the lower transition according to whether 1k D Ak Bk is smaller or larger than zero, respectively. Note that j1k j is a reliability measure of the selection of a certain sequence as survivor sequence. 0 In other words, if j1k j is small, there is a non-negligible probability that the bit bk 2
!sk
associated with the transition from state s.i / 1 to sk on the survivor sequence is in error. k The difference j1k j D k yields the value of the soft decision for bk 2 , in case the nal sequence chosen by the Viterbi algorithm includes the state sk ; conversely, this information is disregarded. Thus the DMA can be formulated as follows. For each state sk of the trellis diagram at instant k, the metric 0k .sk / and the most 0 0 recent .K d C 1/ bits of the survivor sequence b0 .sk / D fbk ; : : : ; bk K d g are memorized, where K d denotes the path memory depth of the VA. Furthermore, the reliability measures .sk / D fk ; : : : ; k K d g associated with the bits b0 .sk / are also memorized. Interpreting O bk and bk as binary symbols in the alphabet f 1; 1g (see note 6 on page 915), the soft output associated with bk is given by Q O `k D bk kC2 O where fbk g is the sequence of information bits associated with the ML sequence. (11.299)
924
The soft-output VA (SOVA). As for the DMA, the SOVA determines the difference between the metrics of the survivor sequences on the paths that converge to each state of the trellis, and updates at every instant k the reliability information .sk / for each state of the trellis. To perform this update, the sequences on the paths that converge to a certain state are compared to identify the positions at which the information bits of the two sequences differ. With reference to Figure 11.24, we denote the two paths that converge to the state (00) at instant k as path 0 and path 1. Without loss of generality we assume that the sequence associated with path 0 is the survivor sequence, and thus the sequence with the smaller cost; furthermore we dene .s.0/ / D f.0/ ; : : : ; .0/ K d g and .s.1/ / D f.1/ ; : : : ; .1/ K d g k k k k k k as the two reliability vectors associated with the information bits of two sequences. If one information bit along path 0 differs from the corresponding information bit along path 1, then its reliability is updated according to the rule for i D k Kd ; : : : ; k 1 if bi.0/2 6D bi.1/2 (11.300)
.0/ i D min.j1k j; i /
With reference to Figure 11.24, the two sequences on path 0 and on path 1 diverge from state sk D .00/ at instant k 4. The two sequences differ in the associated information bits at the instants k and k 1; therefore, only k 1 will be updated. Modied SOVA. In the modied version of the SOVA, the reliability of an information bit along the survivor path is also updated if the information bit is the same, according to the rule for i D k ( i D Kd ; : : : ; k 1 if bi.0/2 6D bi.1/2 if bi.0/2 D bi.1/2 (11.301)
Note that (11.300) is still used to update the reliability if the information bits differ; this version of the SOVA gives a better estimate of i . As proved in [19], if the VA is used as decoder, the modied SOVA is equivalent to Max-Log-MAP decoding. An example of how the modied SOVA works is illustrated in Figure 11.25.
11.5
Turbo codes
Turbo codes, proposed in 1993 by Berrou and Glavieux [20, 21], constitute an evolution of concatenated codes, in the form of parallel concatenated convolutional codes (PCCC), and allow reliable transmission of information at rates near the Shannon limit [20, 21, 22]. As will be discussed in this section, the term turbo, even though it is used to qualify these codes, is rather tied to the decoder, whose principle is reminiscent of that of turbo engines.
Encoding
For the description of turbo codes, we refer to the rst code of this class that appeared in the scientic literature [20, 21]. A sequence of information bits is encoded by a simple
925
recursive systematic convolutional (RSC) binary encoder with code rate 1=2, to produce a rst sequence of parity bits, as illustrated in Figure 11.26. The same sequence of information bits is permuted by a long interleaver and then encoded by a second recursive systematic convolutional encoder with code rate 1=2 to produce a second sequence of parity bits. Then the sequence of information bits and the two sequences of parity bits are transmitted. Note
926
bk
(1) ck (2) ck
interleaver
ck(3)
that the resulting code has rate Rc D 1=3. Higher code rates Rc are obtained by transmitting only some of the parity bits (puncturing). For example, for the turbo code in [20, 21], a code rate equal to 1/2 is obtained by transmitting only the bits of the parity sequence 1 with odd indices, and the bits of the parity sequence 2 with even indices. A specic example of turbo encoder is reported in Figure 11.27. The exceptional performance of turbo codes is due to one particular characteristic. We can think of a turbo code as being a block code for which an input word has a length equal to the interleaver length, and a code word is generated by initializing to zero the memory elements of the convolutional encoders before the arrival of each input word. This block code has a group structure. As for the usual block codes, the asymptotic performance, for large values of the signal-to-noise ratio, is determined by the code words of minimum weight and by their number. For low values of the signalto-noise ratio, also the code words of non-minimum weight and their multiplicity need to be taken into account. Before the introduction of turbo codes, the focus on designing codes was mainly on asymptotic performance, and thus on maximizing the minimum distance. With turbo codes, the approach is different. Because of the large ensemble of code words, the performance curve, in terms of bit error probability as a function of the signal-to-noise ratio, rapidly decreases for low values of the signal-to-noise ratio. For Pbit lower than 10 5 , where performance is determined by the minimum distance between code words, the bit error probability curve usually exhibits a reduction in the value of slope. The two encoders that compose the scheme of Figure 11.26 are called component encoders and they are usually identical. As mentioned above, Berrou and Glavieux proposed two recursive systematic convolutional encoders as component encoders. Later it was shown that it is not necessary to use systematic encoders [23, 17]. Recursive convolutional codes are characterized by the property that the code bits at a given instant do not depend only on the information bits at the present instant and the
927
previous instants, where is the constraint length of the code, but on all the previous information bits, as the encoder exhibits a structure with feedback. Starting from a non-recursive nonsystematic convolutional encoder for a code with rate 1=n 0 , it is possible to obtain in a very simple way a recursive systematic encoder for a code with the same rate and the same code words, and hence with the same free distance H dfree . Obviously, for a given input word, the output code words will be different in the two cases. Consider for example a non-recursive nonsystematic convolutional encoder for a code with code rate 1=2. The code bits can be expressed in terms of the information bits as (see (11.254)) c .1/ .D/ D g .1;1/ .D/ b.D/ 0 c .2/ .D/ D g .2;1/ .D/ b.D/
0
(11.302)
The corresponding recursive systematic encoder is obtained by dividing the polynomials in (11.302) by g .1;1/ .D/, and implementing the functions c.1/ .D/ D b.D/ g .2;1/ .D/ b.D/ c.2/ .D/ D .1;1/ g .D/ Let us dene d.D/ D b.D/ c.2/ .D/ D .1;1/ .2;1/ .D/ g g .D/ (11.305) (11.303) (11.304)
then the code bits can be expressed as a function of the information bits and the bits of the sequence fdk g as
.1/ c k D bk X .2;1/ .2/ gi dk ck D i D0 .1;1/ where, using the fact that g0 .D/ D 1, from (11.305) we get
(11.306)
i
(11.307)
d k D bk C
X i D1
gi.1;1/ dk
(11.308)
We recall that the operations in the above equations are in GF(2). Another recursive systematic encoder that generates a code with the same free distance is obtained by exchanging the role of the polynomials g .1;1/ .D/ and g .2;1/ .D/ in the above equations. One recursive systematic encoder corresponding to the non-recursive nonsystematic encoder of Figure 11.9(a) is illustrated in Figure 11.28. The 16-state component encoder for a code with code rate 1=2 used in the turbo code of Berrou and Glavieux [20, 21], is shown in Figure 11.29. The encoder in Figure 11.27, with an 8-state component encoder for a code with code rate 1/2, is adopted in the standard for third generation universal mobile telecommunications systems (UMTS) [24]. Turbo codes
928
(1) ck
bk
dk
dk1
dk2
(2) ck
Figure 11.28. Recursive systematic encoder that generates a code with the same free distance as the non-recursive nonsystematic encoder of Figure 11.9(a).
c (1) k
bk
(2) ck
Figure 11.29. A 16-state component encoder for the turbo code of Berrou and Glavieux.
are also used in digital video broadcasting (DVB) [25] standards and in space telemetry applications as dened by the Consultative Committee for Space Data Systems (CCSDS) [26]. In [27] are listed generator polynomials of recursive systematic convolutional encoders for codes with rates 1/2, 1/3, 1/4, 2/3, 3/4, 4/5, and 2/4, that can be used for the construction of turbo codes. Another fundamental component in the structure of turbo codes is represented by a nonuniform interleaver. We recall that a uniform8 interleaver, as that described in Section 11.3.2, operates by writing input bits in a matrix by rows and reading them by columns. In practice, a non-uniform interleaver determines the permutation of the sequence of input bits so that adjacent bits in the input sequence are separated, after the permutation, by a number of bits that varies with the position of the bits in the input sequence. The interleaver determines directly the minimum distance of the code and therefore performance for high values of the signal-to-noise ratio. Nevertheless, the choice of the interleaver is not critical for low values of the signal-to-noise ratio. Beginning with the interleaver originally proposed
8
The adjective uniform, referred to an interleaver, is used with a different meaning in [23].
929
in [20, 21], various interleavers have since been proposed (see [28] and references contained therein). One of the interleavers that yields better performance is the so-called spread interleaver [29]. Consider a block of M1 input bits. The integer numbers that indicate the position of these bits after the permutation are randomly generated with the following constraint: each integer randomly generated is compared with the S1 integers previously generated; if the distance from them is shorter than a prexed threshold S2 , the generated integer is discarded and another one is generated until the condition is satised. The two parameters S1 and S2 must be larger than the memory of the two-component encoders. If the two-component encoders are equal, it is convenient to choose S1 D S2 . The computation time needed to generate the interleaver increases with S1 and S2 , and there is no guarantee that the procedure terminates successfully. Empirically, it has been veried that, p choosing both S1 and S2 equal to the closest integer to M1 =2, it is possible to generate the interleaver in a nite number of steps. Many variations of the basic idea of turbo codes have been proposed. For example, codes generated by serial concatenation of two convolutional encoders, connected by means of a non-uniform interleaver [30]. Parallel and serial concatenation schemes were then extended to the case of multilevel constellations to obtain coded modulation schemes with high spectral efciency (see [31] and references contained therein).
930
Figure 11.30. Principle of the decoder for a turbo code with rate 1=3.
The algorithms for iterative decoding introduced with the turbo codes were also immediately applied in wider contexts. In fact, this iterative procedure may be used every time the transmission system includes multiple processing elements with memory interconnected by an interleaver. Iterative decoding procedures may be used, for example, for detection in the presence of intersymbol interference, also called turbo equalization or turbo detection [32] (see Section 11.6), for non-coherent decoding [33, 34], and for joint detection and decoding in the case of transmission over channels with fading [35]. Before discussing in detail iterative decoding, it is useful to revisit the FBA.
k D 0; 1; : : : K
(11.309)
where k0 can be seen either as the number of encoder inputs or as the length of an information vector. As the convolutional encoder is systematic, at instant k the state of the
931
; : : : ; dk
(11.310)
which has a number of components N2 D .n 0 k0 /, equal to the number of the encoder memory elements. The set of states S, that is the possible values assumed by sk , is given by sk 2 S D f 1 ; 2 ; : : : ; Ns g (11.311)
where Ns D 2 N2 is the number of encoder states. It is important to observe that the convolutional encoder can be seen as a sequential nite-state machine with i.i.d. input bk , and state transition function sk D f s .bk ; sk 1 /. Hence, for a given information vector bk , the transition from state sk 1 D i to state sk D j is unique, in correspondence of which a code vector is generated, that is expressed as
.1/ .2/ ck D [ck ; ck ; : : : ; ck 0 ; ck .k / .k0 C1/
; : : : ; ck
.n 0 /
] D [ c.s/ ; ck ] k
. p/
(11.312)
where the superscript .s/ denotes systematic bits, and . p/ denotes parity check bits. Then . p/ c.s/ D bk , and from (11.307) we can express ck as a function of sk and sk 1 as k ck
. p/
D f . p/ .sk ; sk
1/
(11.313)
The values assumed by the code vectors are indicated by D [ .1/ ; .2/ ; : : : ; .k0 / ; .k0 C1/ ; : : : ; .n 0 / ] D [ .s/ ; . p/ ] (11.314)
For simplicity, we assume that the code binary symbols so determined are transmitted by a binary modulation scheme over an AWGN channel. In this case, at the decision point of the receiver, we get the signal (see (11.275)), z k D ck C wk (11.315)
where ck 2 f 1; 1g denotes a code bit, and fwk g is a sequence of real-valued i.i.d. Gaussian noise samples with variance I2 . It is useful to organize the samples fz k g into subsequences that follow the structure of the code vectors (11.312). Then we introduce the vectors
.1/ .2/ zk D [ z k ; z k ; : : : ; z k 0 ; z k .k / .k0 C1/
; : : : ; zk
.n 0 /
] D [ z.s/ ; zk ] k
. p/
. p/
(11.316)
(11.317)
We recall from Section 8.10 that the FBA yields the detection of the single information vector bk , k D 0; 1; : : : ; K 1, expressed as O O .1/ O .2/ O .k / bk D [ bk ; bk ; : : : ; bk 0 ] (11.318)
through the computation of the a posteriori probability. We also recall that in general for a sequence a D [a0 ; : : : ; ak ; : : : ; a K 1 ], with the notation alm we indicate the subsequence formed by the components [al ; alC1 ; : : : ; am ].
932
(11.319)
k D 0; 1; : : : ; K
(11.320)
We note that the likelihood associated with the individual bits of the information vector bk are obtained by suitably adding (11.319), as
.i K K Lk;i ./ D P[bk / D j z0 1 D 0 1 ] X Lk . .s/ / 2 f 1; 1g D .s/ 2f 1;1gk0 .i/ D
(11.321)
In a manner similar to the analysis of page 668, we introduce the following quantities: 1. The state transition probability 5. j j i/ D P[sk D j j sk 1 D i ], that assumes non-zero values only if there is a transition from the state sk 1 D i to the state sk D j for a certain input .s/ , and we write 5. j j i/ D P[bk D .s/ ] D L.a/ . .s/ / k (11.322)
L.a/ . .s/ / is called the a priori information on the information vector k bk D .s/ , and is one of the soft inputs. 2. For an AWGN channel the channel transition probability pzk . k j j; i/ can be separated into two contributions, one due to the systematic bits and the other to the parity check bits, pzk . k j j; i/ D P[zk D k j sk D j ; sk
1
D i]
1 1
D i] D i]
. p/ . p/
D k
. p/
j sk D j ; sk
j ck
. p/
D . p/ ] (11.323)
B D @@ q 00
1 2 I2
A 1 A
1n 0
k0
B@ 1 A @ q 2 I2
1 . p/ jj . p/ jj2 C 2 I2 k
933
3. We merge (11.322) and (11.323) into one variable (see (8.229)), C k . j j i/ D P[zk D k ; sk D j j sk 1 D i ] D pzk . k j j; i/ 5. j j i/ 0 1n 0 1 A . p/ .s/ D @q C k . j j i/ Ck . j j i/ 2 2 I where
.s/ Ck . j . p/
(11.324)
j i/ D e
L.a/ . .s/ / k
(11.325) (11.326)
C k . j j i/ D e
The two previous quantities are related, respectively, to the systematic bits and the parity check bits of a code vector. Observe that the exponential term in (11.325) represents the reliability of a certain a priori information L.a/ . .s/ / associated k with .s/ . 4. The computation of the forward and backward metrics is carried out as in the general case. - Forward metric, for k D 0; 1; : : : ; K Fk . j/ D
Ns X `D1
1:
1 .`/
C k . j j `/ Fk 1; K
j D 1; : : : ; Ns
(11.327)
2; : : : ; 0: i D 1; : : : ; Ns (11.328)
Suitable initializations are obtained, respectively, through (8.237) and (8.244). 5. By using the total probability theorem, the likelihood (11.319) can be written as Lk . .s/ / D A
Ns X
P[sk
K D i ; s k D j ; z0
K D 0
] (11.329)
i D1 j D f s . .s/ ; i / where f s is the state transition function, and the multiplicative constant A D K K 1=P[z0 1 D 0 1 ] is irrelevant for vector detection as can be seen from (11.320). We note that the summation in (11.329) is over all transitions from the general state sk 1 D i to the state sk D j D f s . .s/ ; i / generated by the information vector bk D .s/ . On the other hand, the probability in (11.329) can be
934
written as P[sk
1 K D i ; s k D j ; z0 1 K D 0 1 1
K K D P[zkC11 D kC11 j sk
k k D i ; s k D j ; z0 D 0 ] 1 k D i ; z0 1 k D 0 1
P[sk D j ; zk D k j sk
k P[sk 1 D i ; z0 1 D K K P[zkC11 D kC11 j sk
] (11.330)
k 0 1]
D j]
1
P[sk D j ; zk D k j sk D Bk . j/ Ck . j j i/ Fk
1 .i/
D i ] P[sk
k D i ; z0
k D 0
Substituting for C k . j j i/ the expression in (11.324), the likelihood becomes Lk . .s/ / D B Lk . .s/ / Lk
.a/ .int/
. .s/ / Lk
.ext/
. .s/ /
(11.331)
(11.332)
Bk . j/ Ck . j j i/ Fk
. p/
1 .i/
(11.333)
i D1 j D f s . .s/ ; i / Observing each term in (11.331), we make the following considerations. i. L.a/ . .s/ / represents the a priori information on the information vector bk D .s/ . k ii. L.int/ . .s/ / depends on the received samples associated with the information veck tor and on the channel characteristics. iii. L.ext/ . .s/ / represents the extrinsic information extracted from the received samk ples associated with the parity check bits. This is the incremental information on the information vector obtained by the decoding process. 6. Typically it is easier to work with the logarithm of the various likelihoods. We associate with each bit of the code vector ck a log-likelihood ratio (LLR) that depends on the channel (see (11.285)), that is
.in/ k
D [ `.in;1/ ; : : : ; `.in;n 0 / ] D [ k k
.in;s/ .in; p/ ; k ] k
(11.334)
935
2 k I2
(11.335)
where k is the observation at the instant k. We dene now two quantities that are related, respectively, to the systematic bits and the parity check bits of the code vector, as `.s/ . .s/ / D k and `k . j; i/ D
. p/ n0 1 X .in;m/ .m/ ` 2 mDk C1 k
0
(11.336)
(11.337)
Expressing (11.325) and (11.326) as a function of the likelihoods (11.336) and (11.337), apart from factors that do not depend on f .m/ g; m D 1; : : : ; n 0 , we get C k.s/ . j j i/ D e`k
0 .s/
. .s/ / `k . .s/ /
.a/
(11.339)
and Ck
0 . p/
. j j i/ D e`k
. p/
. j;i /
(11.340)
To compute the forward and backward metrics, we use, respectively, (11.327) and 0 . p/ 0 0 (11.328), where the variable C k . j j i/ is replaced by Ck . j j i/ D Ck.s/ . j j i/ Ck . j j i/. Similarly in (11.333) C k . j j i/ is replaced by Ck . j j i/. Taking the logarithm of (11.333) we obtain the extrinsic component `.ext/ . .s/ /. k Finally, from (11.331), by ignoring non-essential terms, the log-likelihood associated with the information vector bk D .s/ is given by `k . .s/ / D `.a/ . .s/ / C `.int/ . .s/ / C `.ext/ . .s/ / k k k where `.int/ . .s/ / D `.s/ . .s/ / is usually called the intrinsic component. k k Expression (11.341) suggests an alternative method to (11.333) to obtain .ext/ `k . .s/ /, which uses the direct computation of `k . .s/ / by (11.329) and (11.330), 0 where C k . j j i/ is replaced by Ck . j j i/, whose factors are given in (11.339) and (11.341)
. p/
0 . p/
936
(11.340). From the known a priori information `.a/ . .s/ / and the intrinsic information k (11.336), from (11.341) we get `.ext/ . .s/ / D `k . .s/ / k `.a/ . .s/ / k `.int/ . .s/ / k (11.342)
Going back to the expression (11.341), detection of the vector bk is performed according to the rule O bk D arg max `k . .s/ /
.s/
(11.343)
Note that to compute `.ext/ . .s/ / (`k . .s/ /) by the logarithm of (11.333) (or (11.329) and k (11.330)), we can use the Max-Log-MAP method discussed in Section 8.10.2. Example 11.5.1 (Systematic convolutional code with rate 1=2) For a convolutional code with rate Rc D 1=2 the information vector bk D [bk ] is composed .s/ . p/ of only one bit (k0 D 1), like the systematic part and the parity check part of ck D [ck ; ck ]. In this case it is sufcient to determine the log-likelihoods `k . 1/ and `k .1/, or better the LLR Lk .1/ D `k .1/ `k . 1/ (11.344) `k D ln Lk . 1/ Detection of the information bit is performed according to the rule O bk D sgn.`k / The a priori information at the decoder input is given by `k D ln
.a/
(11.345)
P[bk D 1] P[bk D 1]
(11.346)
]De
`.a/ . .s/ / k
.s/ `.a/ k
.s/ `.a/ k
1 .a/ 2 `k `.a/ k
e2
.s/ `.a/ k
.s/ D f 1; 1g (11.347)
1Ce
(11.348)
.s/ .`k 1
.in;s/
C`k /
.a/
(11.349) (11.350)
Ck
0 . p/
. j j i/ D e 2
. p/ `k
.in; p/
937
The extrinsic component is obtained starting from (11.333) and using the above variables `.ext/ k D ln L.ext/ . k L.ext/ .1/ k 1/ D `.ext/ .1/ k `.ext/ . 1/ k (11.351)
From (11.341), apart from irrelevant terms, the LLR associated with the information bit bk can be written as `k D `.a/ C `.int/ C `.ext/ k k k where the meaning of each of the three contributions is as follows. - A priori information `.a/ . It is an a priori reliability measure on the bit bk . This value k can be extracted either from the known statistic of the information sequence or, in the case of iterative decoding of turbo codes, from the previous analysis. - Channel information `.int/ D `.in;s/ . As it is evident from the case of binary moduk k .s/ lation, where `.in;s/ D 22 k , if the noise variance is low, the contribution of `.int/ k k usually dominates with respect to the other two terms; in this case bit detection is .s/ simply obtained by the sign of k . - Extrinsic information `.ext/ . It is a reliability measure that is determined by the k redundancy in the transmitted sequence. This contribution improves the reliability of transmission over a noisy channel using the parity check bits. The decomposition (11.352) forms the basis for the iterative decoding of turbo codes. Observation 11.5 In the case of multilevel modulation and/or for transmission over channels with ISI, the previous formulation of the decoding scheme remains unchanged, provided the expression (11.285) for f`.in;m/ g; m D 1; : : : ; n 0 , is used in place of (11.335). k Example 11.5.2 (Nonsystematic code and LLR associated with the code bits) Consider the case of a nonsystematic code. If the code is also non-recursive, for example as illustrated on page 915 for k0 D 1, we need to use in place of (11.310) the state denition (11.273). Now all bits are parity check bits and (11.312) and (11.316) become, respectively, ck D ck zk D
. p/ .1/ D [ ck ; : : : ; c k .n 0 / I
(11.352)
(11.353) (11.354)
. p/ zk
.1/ .n [ zk ; : : : ; zk 0/ ]
.1/ .k However, the information vector is still given by bk D [ bk ; : : : ; bk 0 / ] with values D [ .1/ ; : : : ; .k0 / ]; .i / 2 f 1; 1g. The likelihood (11.319) is given by K Lk ./ D P[bk D j z0 1 K D 0 1
(11.355)
938
The various terms with superscript .s/ of the previous analysis vanish by setting k0 D 0. Therefore (11.336) and (11.337) become `.int/ . .s/ / D `.s/ . .s/ / D 0 k k and `k . j; i/ D
. p/ n0 1 X .in;m/ .m/ ` 2 mD1 k
(11.356)
(11.357)
where D . p/ D [ .1/ ; : : : ; .n 0 / ] D f . j ; i / is the code vector associated with the transition from state i to state j . Note that, apart from irrelevant factors, (11.357) coincides with (11.291). For k0 D 1, it is convenient to use LLRs; in particular (11.352) yields a LLR associated with the information bit bk that is given by `k D `.a/ C `.ext/ k k (11.358)
where `.ext/ can be obtained directly using (11.351), (11.340), and (11.333). k .q/ In some applications it is useful to associate a LLR with the encoded bit ck ; q D 1; : : : ; n 0 , rather than to the information bit bk . We dene N `k;q D ln
.q/ P[ck
P[ck
.q/
K D 1 j z0
K D 0
1j
K z0 1
K 0 1]
(11.359)
N Let `.a/ be the a priori information on the code bits. The analysis is similar to the previous k;q case but now, with respect to the encoder output, ck is regarded as an information bit, .m/ while the remaining bits ck ; m D 1; : : : ; n 0 ; m 6D q, are regarded as parity check bits. Equations (11.336), (11.337), (11.349), and (11.350), are modied as follows: 1 .in;q/ .q/ N `.s/ . .q/ / D `k k;q 2 n0 1 X .in;m/ .m/ N. p/ ` `k;q . j; i/ D 2 mD1 k
m 6D q
.q/
(11.360) (11.361)
and
.s/ C k;q . j j i/ D e 2
0
.q/ .`k
.in;q/
N C`.a/ / k;q
(11.362) (11.363)
C k;q . j j i/ D e
0 . p/
N. p/ `k;q . j;i /
N Associated with (11.363) we obtain `.ext/ by using (11.351) and (11.333). The overall result k;q is given by
.in;q/ N N N `k;q D `.a/ C `k C `.ext/ k;q k;q
q D 1; : : : ; n 0
(11.364)
939
Example 11.5.3 (Systematic code and LLR associated with the code bits) With reference to the previous example, if the code is systematic, whereas (11.352) holds
.1/ for the systematic bit ck , for the parity check bits ck the following relations hold [37]. For k0 D 1 let be the value of the information bit bk , bk D , with 2 f 1; 1g, associated with the code vector .q/
ck D D [; .2/ ; : : : ; .n 0 / ] where we assume .1/ D . For q D 2; : : : ; n 0 , we get 1 .in;q/ .q/ N `.s/ . .q/ / D `k k;q 2
n0 1 X .in;m/ .m/ 1 .a/ N. p/ ` C `k `k;q . j; i/ D 2 mD1 k 2
m 6D q
(11.365)
(11.366)
(11.367)
where `k
.a/
.q/ `k
.in;q/
(11.368)
N. p/
(11.369)
N From (11.369) we get `.ext/ using (11.351) and (11.333). The overall result is given by k;q
.in;q/ N N C `.ext/ `k;q D `k k;q
(11.370)
Iterative decoding
In this section we consider the iterative decoding of turbo codes with k0 D 1. In this case, as seen in Example 11.5.1, using the LLRs simplies the procedure. In general, for k0 > 1 we should refer to the formulation (11.341). We now give a step-by-step description of the decoding procedure of a turbo code with rate 1=3, of the type shown in Figure 11.27, where each of the two component decoders DEC1 and DEC2 implements the FBA for recursive systematic convolutional codes with rate 1=2. The decoder scheme is shown in Figure 11.30 where the subscript in LLR corresponds to the component decoder. In correspondence to the information bit bk , the turbo code generates the vector
.1/ .2/ .3/ ck D [ck ; ck ; ck ]
(11.371)
.in/ k
.1/ where ck D bk . We now introduce the following notation for the observation vector that relates to the considered decoder: .in/ k
D [`.in;s/ ; `k;1 k
.in; p/
; `k;2
.in; p/
(11.372)
940
where `.in;s/ corresponds to the systematic part, and `k;1 and `k;2 correspond to the k parity check parts generated by the rst and second convolutional encoder, respectively. If some parity check bits are punctured to increase the rate of the code, at the receiver the corresponding LLRs `.in;m/ are set to zero. k 1. First iteration 1.1 Decoder DEC1 If the statistic of the information bits is unknown, then the bits of the information sequence are considered i.i.d. and the a priori information is zero, `.a/ D ln k;1 For k D 0; 1; 2; : : : ; K P[bk D 1] D0 P[bk D 1]
.in; p/
0 . p/ 0 .s/
.in; p/
.in; p/
to (11.349) and (11.350) the variables C k and C k , and from these the corresponding forward metric Fk . j/ (11.327). After the entire sequence has been received, we compute the backward metric Bk .i/ (11.328) and, using (11.333), we nd L.ext/ .1/ and L.ext/ . 1/. The decoder soft output is the extrinsic infork;1 k;1 mation obtained by the LLR `.ext/ D ln k;1 L.ext/ .1/ k;1 L.ext/ . 1/ k;1 (11.374)
1.2 Interleaver .3/ Because of the presence of the interleaver, the parity check bit cn is obtained in correspondence to a transition of the convolutional encoder state determined by the information bit bn , where n depends on the interleaver pattern. In decoding, the extrinsic information `.ext/ , extracted from DEC1 , and the systematic obserk;1 vation `.in;s/ are scrambled by the turbo code interleaver and associated with k .in; p/ the corresponding observation `n;2 to form the input of the second component decoder.
1.3 Decoder DEC2 The extrinsic information generated by DEC1 is set as the a priori information .a/ `n;2 to the component decoder DEC2 ,
.a/ `n;2 D ln
(11.375)
.ext/ The basic idea consists in supplying DEC2 only with the extrinsic part `n;1 of `n;1 , in order to minimize the correlation between the a priori information and the observations used by DEC2 . Ideally, the a priori information should be an independent estimate. .ext/ As done for DEC1 , we extract the extrinsic information `n;2 .
941
1.4 Deinterleaver The deinterleaver realizes the inverse function of the interleaver, .ext/ so that the extrinsic information extracted from DEC2 , `n;2 , is synchronized of DEC1 . By a feedback loop the a posteriori information `.ext/ is placed at k;2 the input of DEC1 as a priori information `.a/ , giving origin to an iterative k;1 structure. with the systematic part `.in;s/ and the parity check part `k;1 k
.in; p/
of the observation
2. Successive iterations Starting from the second iteration each component decoder has at its input an a priori information. The information on the bits become more reliable as the a priori information stabilizes in sign and increases in amplitude. 3. Last iteration When the decoder achieves convergence, the iterative process can stop and form the overall LLR (11.352), `k;overall D `.in;s/ C `.ext/ C `.ext/ k k;1 k;2 k D 0; 1; : : : ; K 1 (11.376)
To make decoding more reliable, the nal state of each component decoder is set to zero, thus enabling an initialization of the backward metric as in (8.244). As illustrated in Figure 11.31, at the instant following the input of the last information bit, that is for k D K , the commutator is switched to the lower position, and therefore we have dk D 0; .1/ .2/ after clock intervals the zero state is reached. The bits ck and ck , for k D K ; K C 1; : : : ; K C 1, are appended at the end of the code sequence to be transmitted.
Performance evaluation
Performance of the turbo code with the encoder of Figure 11.27 is evaluated in terms of error probability and convergence of the iterative decoder implemented by the FBA. For the memoryless AWGN channel, error probability curves versus E b =N0 are plotted in Figure 11.32 for a sequence of information bits of length K D 640, and various numbers of
ck(1)
(2) ck
bk
dk
dk1
dk2
942
10
10
10
10
P(dec)
bit
10
10
10
10 0.25
1 2 3 4 6 8 0 0.25 0.5
Eb/N0 (dB)
0.75
1.25
1.5
1.75
Figure 11.32. Performance of the turbo code dened by the UMTS standard, with length of the information sequence K D 640, and various numbers of iterations of the iterative decoding process.
10
10
10
P(dec)
bit
10
10
10
10
11
12
Number of iterations
Figure 11.33. Curves of convergence of the decoder for the turbo code dened by the UMTS standard, for K D 320 and various values of Eb =N0 .
943
10
10
10
10
P(dec) bit
10
4
10
10
10 0.75
0.5
0.25
0.25
Eb/N0 (dB)
0.5
0.75
1.25
1.5
1.75
Figure 11.34. Performance of the turbo code dened by the UMTS standard achieved after 12 iterations, for K D 40, 320 and 640.
iterations of the iterative decoding process. Note that performance improves as the number of iterations increases; however, the gain between consecutive iterations becomes smaller as the number of iterations increases. .dec/ In Figure 11.33, the error probability Pbit is given as a function of the number of iterations, for xed values of E b =N0 , and K D 320. From the behavior of the error probability we deduce possible criteria for stopping the iterative decoding process at convergence [36]. A timely stop of the iterative decoding process leads to a reduction of the decoding delay and of the overall computational complexity of the system. Note, however, that convergence is not always guaranteed. The performance of the code depends on the length K of the information sequence. Figure 11.34 illustrates how the bit error probability decreases by increasing K , for a constant E b =N0 . A higher value of K corresponds to an interleaver on longer sequences and thus the assumption of independence among the inputs of each component decoder is better satised. Moreover, the burst errors introduced by the channel are distributed over all the original sequence, increasing the correction capability of the decoder. As the length of the interleaver grows, also the latency of the system increases.
11.6
We consider the transmitter of Figure 11.35 composed of a convolutional encoder, interleaver, bit mapper and modulator for 16-PAM. Interpreting the channel as a nite-state machine, the overall structure may be interpreted as a serial concatenated convolutional
944
bl
convolutional code
cm
interleaver
cn
S/P
ck
BMAP
ak
modulator
Figure 11.35. Encoder structure, bit mapper, and modulator; for 16-PAM: ck D [c4k ; c4k 1 ; c4k 2 ; c4k 3 ].
`n,det
bit likelihood
(a)
interleaver
`k,det ()
rk
(a,SYM)
(a)
`n,det
(ext)
deinterleaver
`m,dec
(in)
SISO decoder
`m,dec
(ext)
SI decoder
^
bl
code (SCCC). The procedure of SISO detection and SI decoding of page 916 can be made iterative by applying the principles of the previous section, by including a SISO decoding stage. With reference to Figure 11.36, a step by step description follows. 0. Initialization Suppose we have no information on the a priori probability of the code bits, therefore we associate with cn a zero LLR,
.a/ `n;det D 0
(11.378)
1. Detector First we associate a log-likelihood with the two possible values of cn D , 2 f 1; 1g, according to the rule
.a/ .a/ `n;det .1/ D `n;det .a/ `n;det . 1/ D 0
(11.379)
Then we express the symbol ak as a function of the bits fcn g according to the bit mapper, for example, for 16-PAM, ak D BMAP fck D [c4k ; c4k
1 ; c4k 2 ; c4k 3 ]g
(11.380)
945
Assuming the sequence fcn g is a sequence of i.i.d. binary symbols, we associate with each value of the symbol ak the a priori information expressed by the log-likelihood `.a;SY M/ . / D k;det where
3 X tD0
`.a/ 4k
t;det .t /
2A
(11.381)
For multilevel transmission over a channel with ISI, the FBA of Section 8.10.2 provides a log-likelihood for each value of ak . The new feature is that now in (8.224) we take into account the a priori information on the various values of akCL 1 ; then (8.229) becomes C k . j j i/ D q 1 2 I2 e
1 jk u k j2 `.a;SY M/ . / 2 I2 e kCL 1 ;det
(11.382)
where D f . j ; i / 2 A is the symbol associated with the transition from the state i to the state j on the trellis determined by the ISI. If Vk .i/; i D 1; : : : ; Ns , denotes the metric corresponding to the various states of the trellis, we associate with each value of the code bits fcn g the following likelihood: L4.kCL 1 /
t;det ./ D Ns X
i D1 i such that c4.kCL / t D
1
Vk .i/
2 f 1; 1g
t D 0; : : : ; 3 (11.383)
Taking the logarithm of (11.383), we obtain the LLR `n;det D `n;det .1/ `n;det . 1/ (11.384)
To determine the extrinsic information associated with fcn g, we subtract the a priori information from (11.384),
.ext/ `n;det D `n;det .a/ `n;det
(11.385)
Note that in this application, the detector considers the bits fcn g as information bits and the log-likelihood associated with cn at the detector output is due to the channel information9 in addition to the a priori information.
.a/ In [38], the quantity `n;det in (11.385) is weighted by a coefcient, which is initially chosen small, when the a priori information is not reliable, and is increased after each iteration.
2. Deinterleaver .ext/ The metrics `n;det are re-ordered according to the deinterleaver to provide the sequence `.ext/ . m;det
For the iterative decoding of turbo codes, this information is dened as intrinsic.
946
3. Decoder (SISO) As input LLR, we use `.in/ D `.ext/ m;dec m;det and we set `.a/ D 0 m;dec (11.387) (11.386)
in the lack of an a priori information on the code bits fcm g. Indeed, we note that in the various formulae the roles of `.a/ and `.in/ can be interchanged. m;dec m;dec Depending on whether the code is systematic or not, we use the SISO decoding procedure reported in Example 11.5.1 and Example 11.5.2, respectively. In both cases we associate with the encoded bits cm the quantity `.ext/ D `m;dec m;dec `.in/ m;dec (11.388)
that is passed to the SISO detector as a priori information, after reordering by the interleaver. 4. Last iteration After a suitable number of iterations, the various metrics stabilize and from the LLRs f`.in/ g associated with fcm g, the SI decoding of bits fbl g is performed, using the m;dec procedure of Example 11.5.1.
11.7
Low-density parity check (LDPC) codes were introduced by Gallager [6] as a family of linear block codes with parity check matrices containing mostly zeros and only a small number of ones. The sparsity of the parity check matrices dening LDPC codes is the key for the efcient decoding of these codes by a message-passing procedure also known as the sum-product algorithm. LDPC codes and their efcient decoding were reinvented by MacKay and Neal [39, 40] in the mid-1990s, shortly after Berrou and Glavieux introduced the turbo-codes discussed in Section 11.5. Subsequently, LDPC codes have generated interest from a theoretical as well as from a practical viewpoint and many new developments have taken place. It is today well acknowledged that LDPC codes are as good as turbo codes, as they are based on a similar design philosophy. Also the decoding techniques used for both methods can be viewed as different realizations of the same fundamental decoding process. However, the soft input soft output forward backward algorithm of Section 11.5, or suboptimal versions of it, used for turbo decoding is rather complex, whereas the sum-product algorithm used for LDPC decoding lends itself to parallel implementation and is computationally simpler. LDPC codes, on the other hand, may lead to more stringent requirements in terms of storage. Recall that a linear .n 0 ; k0 / block code, where n 0 and k0 denote the transmitted block length and the source block length, respectively, can be described in terms of a parity
947
check matrix H, such that the equation Hc D 0 is satised for all code words c (see (11.20)). Each row of the r0 n 0 parity check matrix, where r0 D n 0 k0 is the number of parity check bits, denes a parity check equation that is satised by each code word c. For example, the (7,4) Hamming code is dened by the following parity check equations 2 3 c1 6 c2 7 2 36 7 c5 D c1 c2 c3 1 1 1 0 1 0 0 6 c3 7 (check 1) 6 7 4 1 1 0 1 0 1 0 5 6 c4 7 D 0 ! c6 D c1 c2 c4 (check 2) (11.389) 6 7 1 0 1 1 0 0 1 6 c5 7 (check 3) c7 D c1 c3 c4 6 7 4 c6 5 c7 LDPC codes differ in major ways with respect to the above simple example; they usually have long block lengths n 0 in order to achieve near Shannon-limit performance, their parity check matrices are dened in nonsystematic form and exhibit a number of ones that is much less than r0 n 0 . A parity check matrix for a .J; K /-regular LDPC code has exactly J ones in each of its columns and K ones in each of its rows. A parity check matrix can generally be represented by a bipartite graph, also called a Tanner graph, with two types of nodes: the bit nodes and the parity check nodes (or check nodes) [41]. A bit node n, representing the code bit cn , is connected to the check node m only if the element .m; n/ of the parity check matrix is equal to 1. No bit (check) node is connected to a bit (check) node. For example, the (7,4) Hamming code can be represented by the graph shown in Figure 11.37. We note in this specic case that, because the parity check matrix is given in systematic form, bit nodes c5 , c6 , and c7 in the associated graph are connected to single distinct check nodes. The parity check matrix of a .J; K /-regular LDPC code leads to a graph where every bit node is connected to precisely J check nodes and every check node is connected to precisely K bit nodes. We emphasize that the performance of an LDPC code depends on the random realization of the parity check matrix H. Hence these codes form a constrained random code ensemble. Graphical representations of LDPC codes are useful for deriving and implementing the iterative decoding procedure introduced in [6]. Gallager decoder is a message-passing
Figure 11.37. Tanner graph corresponding to the parity check matrix of the (7,4) Hamming code.
948
decoder, in a sense to be made clear below, based on the so-called sum-product algorithm, which is a general decoding algorithm for codes dened on graphs.10
Encoding procedure
Encoding is performed by multiplying the vector of k0 information bits b by the generator matrix G of the LDPC code: cT D bT G (11.390)
where the operations are in GF(2). Recall that generator and parity check matrices satisfy the relation HGT D 0 (11.391)
Q Q From (11.27), the parity check matrix in systematic form is H D [A; I], where I is the Q r0 r0 identity matrix, and A is a binary matrix. Recall also that any other r0 n 0 matrix Q H whose rows span the same space as H is a valid parity check matrix. Given the block length n 0 of the transmitted sequence and the block length k0 of the information sequence, we select a column weight J , greater than or equal to 3. To dene the code, we generate a rectangular r0 n 0 matrix H D [A B] at random with exactly J ones per column and, assuming a proper choice of n 0 and k0 , exactly K ones per row. The r0 k0 matrix A and the square r0 r0 matrix B are very sparse. If the rows of H are independent, which is usually true with high probability if J is odd [40], by Gaussian Q elimination and reordering of columns we determine an equivalent parity check matrix H in systematic form. From (11.26), we obtain the generator matrix in systematic form as I I GT D Q D (11.392) B 1A A where I is the k0 k0 identity matrix. Assuming initially antipodal linear signaling over an ideal AWGN channel, for the vector of transmitted symbols a D [a1 ; : : : ; an 0 ]T , ak 2 f 1; 1g, corresponding to the code word c D [c1 ; : : : ; cn 0 ]T , the received vector is given by zDaCw where w denotes a vector of Gaussian noise samples with variance I2 . (11.393)
Decoding algorithm
Adopting the MAP criterion (8.221), the optimal decoder returns the components of the O O O vector b D [b1 ; : : : ; bk0 ] that maximize the a posteriori probabilities O bk D arg max P[bk D j z D ; G]
2f0;1g
k D 1; : : : ; k0
(11.394)
10 A wide variety of other algorithms (e.g., the Viterbi algorithm, the forward backward algorithm, the iterative
turbo decoding algorithm, the fast Fourier transform, : : : ) can also be derived as specic instances of the sum-product algorithm [42].
949
Note that (11.394) is equivalent to the MAP criterion expressed by (11.321). However, an attempt to evaluate (11.394) by the direct computation of the joint probability distribution of the components of b given the observation would be impractical. Assuming the probability of b uniform, and w statistically independent of b, we resort to the knowledge of the parity check matrix to simplify the decoding problem. We will nd the most likely binary vector x such that (see (11.20)) s D Hx D 0 (11.395)
given the received noisy vector z and a valid parity check matrix H. We call checks the elements si ; i D 1; : : : ; r0 , of the vector s, which are represented by the check nodes in the corresponding Tanner graph. Then the aim is to compute the marginal a posteriori probabilities Ln ./ D P[x n D j z D ; s D 0; G] The detected code bits will then be given by cn D arg max Ln ./ O
2f0;1g
2 f0; 1g
n D 1; : : : ; n 0
(11.396)
n D 1; : : : ; n 0
(11.397)
We dene as Hi;n the element with indices .i; n/ of the parity check matrix H. Let L.i/ D fn : Hi;n D 1g; i D 1; : : : ; r0 , be the set of the bit nodes that participate in the check with index i. Also, let L.i/nn be the set L.i/ from which the element with index n has been Q Q removed. Similarly, let M.n/ D fi : Hi;n D 1g; n D 1; : : : ; n 0 , be the set of the check nodes in which the bit with index n participates. The algorithm consists of two alternating steps, illustrated in Figure 11.38, in which quantities qi;n and ri;n , associated with each non-zero element of the matrix H, are iteratively updated. The quantity qi;n denotes the probability that xn D ; 2 f0; 1g, given the information obtained via checks other than check i: qi;n D P[xn D j fsi 0 D 0; i 0 2 M.n/nig; z D ] Moreover, we dene the a posteriori probabilities
qn D P[xn D j s D 0; z D ]
(11.398)
(11.399)
950
Given xn D ; 2 f0; 1g, the quantity ri;n denotes the probability of check i being satised and the other bits having a known distribution (given by the probabilities fqi;n 0 : n 0 2 L.i/nn; 2 f0; 1gg): ri;n D P[si D 0; fxn 0 ; n 0 2 L.i/nng j xn D ; z D ]
(11.400)
In the rst step, the quantities ri;n associated with check node i are updated and passed as messages to the bit nodes checked by check node i. This operation is performed for all check nodes. In the second step, quantities qi;n associated with bit node n are updated and passed as messages to the check nodes that involve bit node n. This operation is performed for all bit nodes. From (11.395), we note the property of (11.400) that
0 ri;n D 1
D1 D1
(11.401)
1
2 1 C e2n = I
1 pn D
1 1Ce
2n = I2
(11.402)
Let qi;n D pn ; n 2 L.i/; i D 1; : : : ; r0 ; 2 f0; 1g. First step. We run through the checks, and for the i-th check we compute for each n 2 L.i/ 0 the probability ri;n that, given xn D 0, si D 0 and the other bits fxn 0 : n 0 6D ng have 0 ; q 1 g. a distribution fqi;n 0 i;n 0 From (11.400) we obtain X 0 P[si D 0 j xn D 0; fxn 0 D n 0 : n 0 2 L.i/nng] ri;n D
n 0 2 f0; 1g : n 0 2 L.i/nn
Y
n 0 2L.i /nn
n qi;n 0
(11.403)
1 Moreover, ri;n D 1 0 ri;n .
The conditional probabilities in the above expression are either one or zero, depending on whether si D 0 or si D 1 is obtained for the hypothesized values of 0 1 xn and fxn 0 g. The probabilities fri;n ; ri;n g can be found efciently by the FBA, as illustrated by the following example.
951
Example 11.7.1 Assume K D 4 and L.i/ D fn 1 ; n 2 ; n 3 ; n 4 g. The observation si can be expressed in terms of the input variables xk , k 2 L.i/, as si D xn 1 C xn 2 C xn 3 C xn 4 D
K X lD1
xnl
(11.404)
x n l D sn k
C xn k
(11.405)
with s 0 D 0, and observe that s n K D si . Following the formulation of the FBA in Section 8.10.2 we dene the quantities: 1. Forward metric: Fn k . j/ D P[s n k D j] From (11.405) we obtain the recursive equation Fn k . j/ D P[xn k D j]Fn k 1 . j/ C P[xn k D j ]Fn k 1 . j/ k D 1; : : : K (11.407) j 2 f0; 1g (11.406)
where j denotes the ones complement of j, with the initial condition Fn 0 .0/ D 1. 2. Backward metric: Bn k . j/ D P[si D 0 j sn k D j] D
1 X mD0
P[si D 0 j sn kC1 D m; sn k D j]P[sn kC1 D m j sn k D j] P[si D 0 j sn kC1 D m]P[x n kC1 D m j] j 2 f0; 1g (11.408)
1 X mD0
using (11.405) and the fact that si is independent of sn k given sn kC1 . From (11.408), we obtain the recursive equation Bn k . j/ D P[xn kC1 D j]Bn kC1 .0/ C P[xn kC1 D j ]Bn kC1 .1/ k D 1; : : : K (11.409)
with the initial condition Bn K C1 .0/ D 1, which is obtained from the observation si D 0. Therefore the probabilities ri;n k , 2 f0; 1g are given by (see (8.244))
0 ri;n k D Fn k .0/Bn k .0/ C Fn k .1/Bn k .1/ 1 and ri;n k D 1 0 ri;n k .
k D 1; : : : ; K
(11.410)
952
0 0 1 Second step. After computing ri;n and ri;n we update the values of the probabilities qi;n 1 and qi;n . From (11.398) we nd
qi;n D
(11.411)
Lumping in i;n the contribution of the terms that do not depend on and using the i.i.d. assumption on the code bits, we obtain qi;n D i;n P[z n D n ; fsi 0 D 0; i 0 2 M.n/nig j xn D ] Y D i;n pn ri 0 ;n
i 0 2M.n/ni
(11.412)
0 1 where i;n is chosen such that qi;n C qi;n D 1. Taking into account the information from all check nodes, from (11.399) we can also compute the pseudo a posteriori 0 1 probabilities qn and qn at this iteration, given by Y 0 0 0 ri;n (11.413) qn D n pn i 2M.n/ 1 qn
1 n pn
1 ri;n
(11.414)
At this point, the algorithm repeats from the rst step. At the end of the second step, one iteration of the decoding algorithm is completed. At each iteration, it is O possible to detect a code word c by the log-MAP criterion (8.277), i.e. detect 1 qn ck D sgn ln 0 O n D 1; : : : ; n 0 (11.415) qn Decoding is stopped if HO D 0, or if some other stopping criterion is met, e.g., c maximum number of iterations is achieved. Messages passed between the nodes need not be probabilities but can be likelihood or log-likelihood ratios. In fact, various simplications of the decoding algorithm have been explored and can be adopted for practical implementations [43, 44]. We note that the sumproduct algorithm for the decoding of LDPC codes has been derived under the assumption that the check nodes si , i D 1; : : : ; r0 , are statistically independent given the bit nodes xn , n D 1; : : : ; n 0 , and vice versa, i.e. the variables of the vectors s and x form a Markov eld [42]. Although this assumption is not strictly true, it turns out that the algorithm yields very good performance with low computational complexity. However, we note that parity check matrices leading to Tanner graphs that exhibit cycles of length four, such as the one depicted in Figure 11.39, should be avoided. In fact, this structure would introduce non-negligible statistical dependence between nodes. In graph theory, the length of the shortest cycle in a graph is referred to as girth. A general method for constructing Tanner graphs with large girth is described in [45].
953
Example of application
We study in this section the application of binary LDPC codes to two-dimensional QAM transmission over an AWGN channel [46]. The block diagrams of the encoding and decoding processes are shown in Figure 11.40. For bit mapping, log2 M code bits are mapped into one QAM symbol taken from an M-point constellation using Gray mapping. At the receiver, the received samples, which represent noisy QAM symbols, are input to a soft detector that provides soft information on individual code bits in the form of a posteriori probabilities. These probabilities are employed to carry out the message-passing LDPC decoding procedure described in the previous section. Assuming that the employed QAM constellation is square, with log2 M equal to an even number, and that the in-phase and quadrature noise components are independent, it is computationally advantageous to perform soft detection independently for the real and imaginary parts of the received complex samples. We will therefore consider only square QAM constellations. Bit mapping for the real or the imaginary part of transmitted QAM ] symbols is performed by mapping a group of 1 log2 M code bits [c0 ; c1 ; : : : ; c 1 2 2 .log2 M/ 1 p that are part of a code word into one of the M real symbols within the set p ADf . M p 1/; . M p 3/; : : : ; 1; C1; : : : ; C. M 1/g (11.416)
Denoting by z n the real or the imaginary part of a noisy received signal, we have z n D an C wn (11.417)
954
where an 2 A, and wn is an AWGN sample with variance I2 . The a posteriori probability that bit c` is zero or one is computed as (see (8.262)) X
2A c` D
.n /2 2 I2
P[c` D j z n D n ] D
X
2A
.n /2 2 I2
1 ` D 0; 1; : : : ; .log2 M/ 2
2 f0; 1g
(11.418) where the summation in the numerator is taken over all symbols an 2 A for which c` D , 2 f0; 1g.
Recall also, from (6.191), that the signal-to-noise ratio per dimension is given by 0 I D 0 D 2R I (11.421)
(11.423)
As illustrated in Figure 6.54, the curve of Pe versus 0 I indicates that the gap to capacity for uncoded QAM with M 1 is equal to 0 gap;d B ' 9:8 dB at a symbol error probability of 10 7 . We therefore determine the value of the normalized signal-to-noise c ratio 0 I needed for the coded system to achieve a symbol error probability of 10 7 , and compute the coding gain at that symbol error probability as G code D 9:8 10 log10 .0 I / dB
c
(11.424)
955
Table 11.22 LDPC codes considered for the simulation and coding gains achieved at a symbol error probability of 10 7 for different QAM constellations. The spectral efciencies are also indicated. k0 Code 1 433 n0 495 code rate k0 =n 0 0:8747 0:8894 0:9358 16-QAM 64-QAM 4096-QAM
4:9 dB 4:6 dB 3:5 dB (3:49 bit/s/Hz) (5:24 bit/s/Hz) (10:46 bit/s/Hz) 6:1 dB 5:9 dB 4:8 dB (3:55 bit/s/Hz) (5:33 bit/s/Hz) (10:62 bit/s/Hz) 6:2 dB 6:1 dB 5:6 dB (3:74 bit/s/Hz) (5:61 bit/s/Hz) (11:22 bit/s/Hz)
From Figure 6.54, as for large signal-to-noise ratios the Shannon limit cannot be approached to within less than 1.53 dB without shaping, we note that an upper limit to the coding gain measured in this manner is about 8:27 dB. Simulation results for three high-rate .n 0 ; k0 / binary LDPC codes are specied in Table 11.22 in terms of the coding gains obtained at a symbol error probability of 10 7 for transmission over an AWGN channel for 16, 64, and 4096-QAM modulation formats. Transmitted QAM symbols are obtained from coded bits via Gray mapping. To measure error probabilities, one code word is decoded using the message-passing (sum-product) algorithm for given maximum number of iterations. Figure 11.41 shows the effect on performance of the maximum number of
Figure 11.41. Performance of LDPC decoding with Code 2 and 16-QAM for various values of the maximum number of iterations.
956
iterations allowed in the decoding process for code 2 specied in Table 11.22 and 16-QAM transmission. The codes given in Table 11.22 are due to MacKay and have been obtained by a random construction method. The results of Table 11.22 indicate that LDPC codes offer net coding gains that are similar to those that have been reported for turbo codes. Furthermore, LDPC codes achieve asymptotically an excellent performance without exhibiting error oors and admit a wide range of trade-offs between performance and decoding complexity.
Bibliography
[1] S. Lin and D. J. Costello Jr., Error control coding. Englewood Cliffs, NJ: PrenticeHall, 1983. [2] R. E. Blahut, Theory and practice of error control codes. Reading, MA: AddisonWesley, 1983. [3] W. W. Peterson and E. J. Weldon Jr., Error-correcting codes. Cambridge, MA: MIT Press, 2nd ed., 1972. [4] J. K. Wolf, Lecture notes. San Diego, CA: University of California. [5] S. B. Wicker and V. K. Bhargava, eds, Reed-Solomon codes and their applications. Piscataway, NJ: IEEE Press, 1994. [6] R. Gallager, Information theory and reliable communication. New York: John Wiley & Sons, 1968. [7] J. Hagenauer and P. Hoeher, A Viterbi algorithm with soft-decision output and its applications, in Proc. GLOBECOM 89, Dallas, Texas, pp. 28282833, Nov. 1989. [8] M. P. C. Fossorier and S. Lin, Soft-decision decoding of linear block codes based on ordered statistics, IEEE Trans. on Information Theory, vol. 41, pp. 13791396, Sept. 1995. [9] M. P. C. Fossorier and S. Lin, Soft-input soft-output decoding of linear block codes based on ordered statistics, in Proc. GLOBECOM 98, Sidney, Australia, pp. 2828 2833, Nov. 1998. [10] D. J. Costello Jr., J. Hagenauer, H. Imai, and S. B. Wicker, Applications of errorcontrol coding, IEEE Trans. on Information Theory, vol. 44, pp. 25312560, Oct. 1998. [11] R. M. Fano, A heuristic discussion on probabilistic decoding, IEEE Trans. on Information Theory, vol. 9, pp. 6474, Apr. 1963. [12] K. Zigangirov, Some sequential decoding procedures, Probl. Peredachi Informatsii, vol. 2, pp. 1325, 1966.
11. Bibliography
957
[13] F. Jelinek, An upper bound on moments of sequential decoding effort, IEEE Trans. on Information Theory, vol. 15, pp. 140149, Jan. 1969. [14] F. Jelinek, Fast sequential decoding algorithm using a stack, IBM Journal of Research and Development, vol. 13, pp. 675685, Nov. 1969. [15] F. Q. Wang and D. J. Costello, Erasure-free sequential decoding of trellis codes, IEEE Trans. on Information Theory, vol. 40, pp. 18031817, Nov. 1994. [16] C. F. Lin and J. B. Andeerson, M-algorithm decoding of channel convolutional codes, in Conf. Rec., Princeton Conf. Inform. Sci. Syst., Princeton, NJ, pp. 362365, Mar. 1986. [17] S. Benedetto and E. Biglieri, Principles of digital transmission with wireless applications. New York: Kluwer Academic Publishers, 1999. [18] G. D. Forney, Jr., Concatenated codes. Cambridge, MA: MIT Press, 1966. [19] M. P. C. Fossorier, F. Burkert, S. Lin, and J. Hagenauer, On the equivalence between SOVA and max-log-MAP decodings, IEEE Communications Letters, vol. 2, pp. 137 139, May 1998. [20] C. Berrou, A. Glavieux, and P. Thitimajshima, Near Shannon limit error-correcting coding and decoding: turbo codes, in Proc. IEEE Int. Conference on Communications, Geneva, Switzerland, pp. 10641070, May 2326, 1993. [21] C. Berrou and A. Glavieux, Near optimum error-correcting coding and decoding: turbo-codes, IEEE Trans. on Communications, vol. 44, pp. 12611271, Oct. 1996. [22] B. Sklar, A primer on turbo code concepts, IEEE Communications Magazine, vol. 35, pp. 94101, Dec. 1997. [23] S. Benedetto and G. Montorsi, Unveiling turbo codes: some results on parallel concatenated coding schemes, IEEE Trans. on Information Theory, vol. 42, pp. 409428, Mar. 1996. [24] 3-rd Generation Partnership Project (3GPP), Technical Specication Group (TSG), Radio Access Network (RAN), Working Group 1 (WG1), Multiplexing and channel coding (TDD), Document TS 1.22, v.2.0.0, Apr. 2000. [25] International Telecommunication Union (ITU), Radiocommunication Study Groups, A guide to digital terrestrial television broadcasting in the VHF/UHF bands, Doc. 11-3/3-E, Mar. 1998. [26] Consultative Committee for Space Data Systems (CCSDS), Telemetry Systems (Panel 1), Telemetry channel coding, Blue Book, CCSDS 101.0-B-4: Issue 4, May 1999. [27] S. Benedetto, R. Garello, and G. Montorsi, A search for good convolutional codes to be used in the construction of turbo codes, IEEE Trans. on Communications, vol. 46, pp. 11011105, Sept. 1998.
958
[28] O. Y. Takeshita and D. J. Costello Jr., New deterministic interleaver designs for turbo codes, IEEE Trans. on Information Theory, vol. 46, pp. 19882006, Sept. 2000. [29] D. Divsalar and F. Pollara, Turbo codes for PCS applications, in Proc. 1995 IEEE Int. Conference on Communications, Seattle, U.S.A., pp. 5459, June 1995. [30] S. Benedetto, D. Divsalar, G. Montorsi, and F. Pollara, Serial concatenation of interleaved codes: performance analysis, design, and iterative decoding, IEEE Trans. on Information Theory, vol. 45, pp. 909926, May 1998. [31] S. Benedetto and G. Montorsi, Versatile bandwidth-efcient parallel and serial turbotrellis-coded modulation, in Proc. 2000 Intern. Symp. on Turbo Codes & Relat. Topics, Brest, France, pp. 201208, Sept. 2000. [32] C. Douillard, M. Jezequel, C. Berrou, A. Picart, P. Didier, and A. Glavieux, Iterative correction of intersymbol interference turbo-equalization, Europ. Trans. on Telecommun. (ETT), vol. 6, pp. 507511, September/October 1995. [33] G. Colavolpe, G. Ferrari, and R. Raheli, Noncoherent iterative (turbo) decoding, IEEE Trans. on Communications, vol. 48, pp. 14881498, Sept. 2000. [34] G. Bauch, H. Khorram, and J. Hagenauer, Iterative equalization and decoding in mobile communications system, in Proc. European Personal Mobile Communications Conference, Bristol, UK, pp. 307312, 1997. [35] P. Hoeher and J. Lodge, Turbo-DPSK: iterative differential PSK demodulation, IEEE Trans. on Communications, vol. 47, pp. 837843, June 1999. [36] J. Hagenauer, E. Offer, and L. Papke, Iterative decoding of binary block and convolutional codes, IEEE Trans. on Information Theory, vol. 42, pp. 429445, Mar. 1996. [37] G. Bauch, H. Khorram, and J. Hagenauer, Iterative equalization and decoding in mobile communications systems, in Proc. EPMCC, pp. 307312, Oct. 1997. [38] A. Picart, P. Didier, and G. Glavieux, Turbo-detection: a new approach to combat channel frequency selectivity, in Proc. IEEE Int. Conference on Communications, pp. 14981502, 1997. [39] D. MacKay and R. Neal, Near Shannon limit performance of low density parity check codes, Electron. Lett., vol. 32, pp. 16451646, Aug. 1996. [40] D. MacKay, Good error-correcting codes based on very sparse matrices, IEEE Trans. on Information Theory, vol. 45, pp. 399431, Mar. 1999. [41] R. M. Tanner, A recursive approach to low complexity codes, IEEE Trans. on Information Theory, vol. 27, pp. 533547, Sept. 1981. [42] B. J. Frey, Graphical models for machine learning and digital communications. Cambridge, MA: MIT Press, 1998.
11. Bibliography
959
[43] M. P. C. Fossorier, M. Mahaljevic, and H. Imai, Reduced complexity iterative decoding of low-density parity check codes based on belief propagation, IEEE Trans. on Computers, vol. 47, pp. 673680, May 1999. [44] X.-Y. Hu, E. Eleftheriou, D. M. Arnold, and A. Dholakia, Efcient implementations of the sum-product algorithm for decoding LDPC codes, in Proc. GLOBECOM 01, San Antonio, TX, Nov. 2001. [45] X.-Y. Hu, E. Eleftheriou, and D. M. Arnold, Progressive edge-growth Tanner graphs, in Proc. GLOBECOM 01, San Antonio, TX, Nov. 2001. R [46] G. Cherubini, E. Eleftheriou, and S. Olcer, On advanced signal processing and coding techniques for digital subscriber lines, Records of the Workshop What is next in xDSL?, Vienna, Austria, Sept. 2000.
960
Appendix 11.A
Assume that code words are sequences of symbols from the nite eld G F.q/ (see Section 11.2.2), all of length n. As there are q n possible sequences, the introduction of redundancy in the transmitted sequences is possible if the number of code words Mc is less than q n . We denote by c a transmitted sequence of n symbols taken from G F.q/. We also assume that the symbols of the received sequence z are from the same alphabet. We dene the error sequence e by the equation (see (11.12) for the binary case) zDcCe (11.425)
where C denotes a component by component addition of the vectors in accordance with the rules of addition in the eld G F.q/. Denition 11.16 The number of non-zero components of a vector x is dened as the weight of the vector, denoted by w.x/. Then w.e/ is equal to the number of errors occurred in transmitting the code word. Denition 11.17 H The minimum distance of a code, denoted dmin , is equal to the minimum Hamming distance between all pairs of code words; i.e. it is the same as for binary codes. We will give without proof the following propositions, similar to those for binary codes on page 830.
H 1. A nonbinary block code with minimum distance dmin can correct all error sequences k j H dmin 1 or less. of weight 2 H 2. A nonbinary block code with minimum distance dmin can detect all error sequences H of weight .dmin 1/ or less. H As in the binary case, we ask for a relation among the parameters of a code: n, Mc , dmin , H and q. It can be proved that for a block code with length n and minimum distance dmin , Mc must satisfy the inequality 6 8 j H k 97 6 dmin 1 =7 6 n < 7 n n n 2 5 H M c 4q 1C .q 1/ C .q 1/2 C C j dmin 1 k .q 1/ ; : 1 2 2
(11.426)
961
H Furthermore, for n and dmin given, it is always possible to nd a code with Mc words, where n n n H n 2 dmin 1 1C .q 1/ C .q 1/ C C .q 1/ Mc D q H 1 2 dmin 1 (11.427)
Linear codes
Denition 11.18 A linear code is a block code with symbols from G F.q/ for which: a) the all zero word is a code word; b) any multiple of a code word is a code word; c) any linear combination of any two code words is a code word. Example 11.A.1 A binary group code is a linear code with symbols from G F.2/. Example 11.A.2 Consider a block code of length 5 having symbols from G F.3/ with code words 0 1 0 2 1 2 0 1 2 0 0 1 0 1 1 2 2 2 0 0 1 0 1 1 2 2 2 0 2 2 1 1 0 1 0 2 0 1 2 2 0 1 1 2 0
(11.428)
It is easily veried that this code is a linear code. We give the following two properties of a linear code.
H 1. The minimum distance of the code, dmin , is given as H c dmin D min w.N /
(11.429)
N where c can be any non-zero code word. Proof. By denition of the Hamming distance between two code words, we get d H .c1 ; c2 / D w.c1 C . c2 // (11.430)
962
By Property b), . c2 / is a code word if c2 is a code word; by Property c), c1 C . c2 / H must also be a code word. As two code words differ in at least dmin positions, there H H is a code word of weight dmin ; if there were a code word of weight less than dmin , H positions. this word would be different from the zero word in fewer than dmin 2. If all code words in a linear code are written as rows of an Mc n matrix, every column is composed of all zeros, or contains all elements of the eld, each repeated Mc =q times.
where the r r matrix B is such that det[B] 6D 0. A generalized nonbinary parity check code is a code composed of all vectors c of length n, with elements from G F.q/, that are the solutions of the equation Hc D 0 The matrix H is called the generalized parity check matrix. Propriety 1 of nonbinary generalized parity check codes. A nonbinary generalized parity check code is a linear code. Proof. a) The all zero word is a code word, as H0 D 0. b) Any multiple of a code word is a code word, because if c is a code word, then Hc D 0. But H.c/ D Hc D 0, and therefore c is a code word; here is any element from G F.q/. c) Any linear combination of any two code words is a code word, because if c1 and c2 are two code words, then H.c1 C c2 / D Hc1 C Hc2 D 0 C 0 D 0, and therefore c1 C c2 is a code word. Property 2 of nonbinary generalized parity check codes. The code words corresponding to the matrix H D [A B] are identical to the code words corresponding to the parity check Q matrix H D [B 1 A; I]. Proof. Same as for the binary case. The matrices in the form [A I] are said to be in canonical or systematic form. Property 3 of nonbinary generalized parity check codes. A code consists of exactly q n q k code words. Proof. Same as for the binary case (see Property 3 on page 834). The rst k D n r symbols are called information symbols, and the last r symbols are called generalized parity check symbols.
r
(11.432)
963
Property 4 of nonbinary generalized parity check codes. A code word of weight w exists if and only if some linear combination of w columns of the matrix H is equal to 0. Proof. c is a code word if and only if Hc D 0. Let c j be the j-th component of c and let hi be the i-th column of H; then if c is a code word we have
n X jD1
hj cj D 0
(11.433)
If c is a code word of weight w, there are exactly w non-zero components of c, say c j1 ; c j2 ; : : : ; c jw ; then c j1 h j1 C c j2 h j2 C C c jw h jw D 0 (11.434)
thus, a linear combination of w columns of H is equal to 0. Conversely, if (11.434) is true, then Hc D 0, where c is a vector of weight w with non-zero components c j1 ; c j2 ; : : : ; c jw . Combining Property 1 of a linear code and Properties 1 and 4 of a nonbinary generalized parity check code, we obtain the following property. Property 5 of nonbinary generalized parity check codes. A code has minimum distance H H dmin if some linear combination of dmin columns of H is equal to 0, but no linear combiH number of columns of H is equal to 0. nation of fewer than dmin Property 5 is fundamental for the design of nonbinary codes. Example 11.A.3 Consider the eld G F.4/, and let be a primitive element of this eld; moreover consider the generalized parity check matrix 1 1 1 1 0 HD (11.435) 1 2 0 1 We nd that no linear combination of two columns is equal to 0. However, there are many linear combinations of three columns that are equal to 0, for example, h1 C h4 C h5 D 0, H h2 C h4 C 2 h5 D 0, ....; hence the minimum distance of this code is dmin D 3.
Acn 1
(11.436)
I A
cn 1
D G T cn 1
(11.437)
The matrix G is called the generator matrix of the code and is expressed as G D [I; AT ] (11.438)
964
so that cT D .cn 1
r T
/ G
(11.439)
Thus the code words, considered as row vectors, are given as all linear combinations of the rows of the matrix G. A nonbinary generalized parity check code can be specied by giving its generalized parity check matrix or its generator matrix. Example 11.A.4 Consider the eld G F.4/ and let be a primitive element of this eld; moreover, consider the generalized parity check matrix (11.435). The generator matrix of this code is given by 2 3 1 0 0 1 1 GD40 1 0 1 5 (11.440) 0 0 1 1 2 There are 64 code words corresponding to all linear combinations of the rows of the matrix G.
D q k code words, c1 ; c2 ; : : : ; cq k .
Step 2: choose any vector, say 2 , that is not a code word; then choose the second set as c1 C 2 ; c2 C 2 ; : : : ; c q k C 2 . Step i: choose any vector, say i , not included in any previous set; choose the i-th set as c1 C i ; c2 C i ; : : : ; c q k C i . The partitioning continues until all q n vectors are used; each set is called a coset, and the vectors i are called coset leaders. The all zero vector is the coset leader for the rst set.
Coset
We give the following properties of the cosets omitting the proofs. 1. Every one of the q n vectors occurs in one and only one coset. 2. Suppose that, instead of choosing i as coset leader of the i-th coset, we choose another element of that coset as the coset leader; then the coset formed by using the new coset leader contains exactly the same vectors as the old coset. 3. There are q r cosets.
965
Syndrome decoding
Another method of decoding is the syndrome decoding. For any generalized parity check matrix H and all vectors z of length n, we dene the syndrome of z, s.z/, as s.z/ D Hz (11.441)
We can show that all vectors in the same coset have the same syndrome and vectors in different cosets have different syndromes. This leads to the following decoding method: Step 100 : compute the syndrome of the received vector, as this syndrome identies the coset in which the received vector is in, and so identies the leader of that coset. Step 200 : subtract the coset leader from the received vector to nd the decoded code word. The difculty with this decoding method is in the second part of step 100 , that is identifying the coset leader that corresponds to the computed syndrome; this step is equivalent to nding a linear combination of the columns of H which is equal to that syndrome, using the smallest number of columns. The algebraic structure of the generalized parity check matrix for certain classes of codes allows for algebraic means of nding the coset leader from the syndrome.
Chapter 12
During the 1980s an evolution in the methods to transmit data over channels with limited bandwidth took place, giving origin to techniques for joint coding and modulation that are generally known by the name of trellis coded modulation (TCM). The main characteristic of TCM lies in the fact that it yields coding gains with respect to conventional modulation techniques without requiring that the channel bandwidth be increased. The rst article on TCM appeared in 1976 by Ungerboeck; later, a more detailed publication by the same author on the principles of TCM [1] spurred considerable interest in this topic [2, 3, 4, 5, 6, 7, 8], leading to a full development of the theory of TCM. TCM techniques use multilevel modulation with a set of signals from a one, two, or multidimensional space. The choice of the signals that generate a code sequence is determined by a nite-state encoder. In TCM, the set of modulation signals is expanded with respect to the set used by an uncoded, i.e. without redundancy, system; in this manner, it is possible to introduce redundancy in the transmitted signal without widening the bandwidth. At the receiver, the signals in the presence of additive noise and channel distortion are decoded by a maximum likelihood sequence decoder. By simple TCM techniques using a four-state encoder, it is possible to obtain a coding gain of 3 dB with respect to conventional uncoded modulation; with more sophisticated TCM techniques, coding gains of 6 dB or more can be achieved (see Chapter 6). Errors in the decoding of the received signal sequence are less likely to occur if the waveforms, which represent the code sequences, are easily distinguishable from each other; in mathematical terms, the signal sequences, represented in the Euclidean multidimensional space, need to be separated by large distances. The novelty of TCM is in postulating the expansion of the set of symbols1 in order to provide the redundancy necessary for the encoding process. The construction of modulation code sequences that are characterized by a free distance, i.e. the minimum Euclidean distance between code sequences, that is much larger than the minimum distance between uncoded modulation symbols, with the same information bit rate, and the same bandwidth and power of the modulated signal, is obtained by the joint design of encoder and bit mapper. The term trellis derives from the similarity between state transition diagrams of a TCM encoder and trellis diagrams
1
In the rst part of this chapter we mainly use the notion of symbols of an alphabet with cardinality M, although the analysis could be conducted by referring to vectors in the signal space as modulation signals. We will use the term signals instead of symbols only in the multidimensional case.
968
of binary convolutional codes; the difference lies in the fact that, in TCM schemes, the branches of the trellis are labeled with modulation symbols rather than binary symbols. Thanks to the use of sophisticated TCM schemes, it was possible to achieve reliable data transmission over telephone channels at rates much higher than 9.6 kbit/s, which for years was considered the practical limit. In the mid-1980s, the rate of 14.4 kbit/s was reached. Transmission at a maximum bit rate of 28.8 kbit/s was later specied in the standard CCITT V.34, and extensions were proposed to achieve the rates of 31.2 kbit/s and 33.6 kbit/s.
12.1
12.1.1
Consider the transmission system illustrated in Figure 6.20, that consists of the modulator, transmission channel, demodulator and data detector. Errors occasionally occur in the symbol detection, and ak 6D ak . Usually the simplest data detector is a threshold detector that O takes an instantaneous hard decision on the value ak of the transmitted symbol, based on the O observation of the sample z k at the demodulator output. Detection is of the nearest-neighbor type, i.e. the detector decides for the symbol of the constellation that is at the minimum Euclidean distance from the received sample z k . The objective of traditional channel coding techniques consists in detecting and/or correcting the errors present in the detected sequence of bits fcm g. Q In the approach followed in Chapter 11, a binary encoder was used to map k0 information binary symbols fb` g in n 0 code binary symbols fcm g. As mentioned in Section 11.1, we note that, if we want to maintain the effective rate 1=Tb of the information message and at the same time the modulation rate 1=T of the system, we need to increase the cardinality of the modulation alphabet. If we do not consider the joint design of encoder and bit mapper, however, a reduction of the bit error probability cannot be efciently achieved, as we see from the following example. Example 12.1.1 Consider an uncoded 4-PSK system and an 8-PSK system that uses a binary error correcting code with rate 2/3; both systems transmit two information bits per modulation interval, which corresponds to a spectral efciency of 2 bit/s/Hz. If the 4-PSK system works with an error probability of 10 5 , for a given signal-to-noise ratio 0, the 8-PSK system works with an error probability larger than 10 2 , due to the smaller Euclidean distance between signals of the 8-PSK system. We must use an error correcting code with minimum Hamming H distance dmin 7 to reduce the error probability to the same value of the uncoded 4-PSK system. A binary convolutional code with rate 2/3 and constraint length 6 has the required H value of dfree D 7. Decoding requires a decoder with 64 states that implements the Viterbi algorithm. However, even after increasing the complexity of the 8-PSK system, we have obtained an error probability only equal to that of the uncoded 4-PSK system. Two problems determine the unsatisfactory result obtained with the traditional approach. The rst is originated by the use of independent hard decisions taken by the detector before
969
Figure 12.1. Block diagram of a transmission system with trellis coded modulation.
decoding; hard input decoding leads to an irreversible loss of information; the remedy is the use of soft decoding (see page 912), whereby the decoder directly operates on the samples at the demodulator output. The second derives from the independent design of encoder and bit mapper. We now consider the transmission system of Figure 12.1, where the transmitted symbol sequence fak g is produced by a nite-state machine having the information bit sequence fb` g as input, possibly with a number of information bits per modulation interval larger than one. We denote by 1=T the modulation rate and by A the alphabet of ak . For an AWGN channel, at the decision point the received samples in the absence of ISI are given by (see (8.173)) z k D a k C wk (12.1)
where fwk g is a sequence of white Gaussian noise samples. Maximum likelihood sequence detection represents the optimum strategy for decoding a sequence transmitted over a dispersive noisy channel. The decision rule consists in determining the sequence fak g closest O to the received sequence in terms of Euclidean distance (see (8.190)) in the set S of all possible code symbol sequences. MLSD is efciently implemented by the Viterbi algorithm, provided that the generation of the code symbol sequences follows the rules of a nite-state machine. In relation to (8.194), we dene as free distance, dfree , the minimum Euclidean distance between two code symbol sequences fk g and fk g, that belong to the set S, given by X 2 jk k j2 fk g; fk g 2 S (12.2) dfree D min
fk g6Dfk g k
The most probable error event is determined by two code symbol sequences of the set S at the minimum distance. The assignment of symbol sequences using a code that is optimized for Hamming distance does not guarantee an acceptable structure in terms of Euclidean distance, as in general the relation between the Hamming distance and the Euclidean distance is not monotonic. Encoder and modulator must then be jointly designed for the purpose of assigning to symbol sequences waveforms that are separated in the Euclidean signal space by a distance
970
equal to at least dfree , where dfree is greater than the minimum distance between the symbols of an uncoded system. At the receiver, the demodulator-decoder does not make errors if the received signal in the Euclidean signal space is at a distance smaller than dfree =2 from the transmitted sequence.
(12.3)
For an input vector bk and a state sk 1 , the rst equation describes the choice of the transmitted symbol ak from a certain constellation, the second the choice of the next state sk . Interdependence between the symbols fak g is introduced without a reduction of the bit rate by increasing the cardinality of the alphabet. For example, for a length K of the sequence of input vectors, if we change A of cardinality M with A0 A of cardinality M 0 > M, and we select M K sequences as a subset of .A0 / K , a better separation of the code sequences in the Euclidean space may be obtained. Hence, we can obtain a minimum distance dfree between any two sequences larger than the minimum distance between signals in A K . Note that this operation may cause an increase in the average symbol energy from E s;u for uncoded transmission to E s;c for coded transmission, and hence a loss in efciency given by E s;c =E s;u . Furthermore, we dene as Nfree the number of sequences that a code sequence has, on average, at the distance dfree in the Euclidean multidimensional space.
Example
Suppose we want to transmit two bits of information per symbol. Instead of using QPSK modulation, we can use the scheme illustrated in Figure 12.2. The scheme has two parts. The rst is a nite-state sequential machine with 8 states, .2/ .1/ .0/ where the state sk is dened by the content of the memory cells sk D [sk ; sk ; sk ]. The
971
Figure 12.2. Eight-state trellis encoder and bit mapper for the transmission of 2 bits per modulation interval by 8-PSK.
.2/ .1/ two bits bk D [bk ; bk ] are input to the FSM, which undergoes a transition from state sk 1 to one of four next possible states, sk , according to the function g. The second part is the bit mapper, which maps the two information bits and one bit that depends on the state, i.e. .2/ .1/ .0/ the three bits [bk ; bk ; sk 1 ], in one of the symbols of an eight-ary constellation according to the function f , for example, an 8-PSK constellation using the map of Figure 12.5. Note that the transmission of two information bits per modulation interval is achieved. Therefore the constellation of the system is expanded by a factor 2 with respect to uncoded QPSK transmission. Recall from the discussion in Section 6.10 that most of the achievable coding gain for transmission over an ideal AWGN channel of two bits per modulation interval can be obtained by doubling the cardinality of the constellation from four to eight symbols. We will see that trellis coded modulation using the simple scheme of Figure 12.2 allows to achieve a coding gain of 3.6 dB. For the graphical representation of the functions f and g, it is convenient to use a trellis diagram; the nodes of the trellis represent the FSM states and the branches represent the possible transitions between states. For a given state sk 1 , a branch is associated with each possible vector bk by the function g, that reaches a next state sk . Each branch is labeled with the corresponding value of the transmitted symbol ak . For the encoder of Figure 12.2 and the map of Figure 12.5, the corresponding trellis is shown in Figure 12.3, where the trellis is terminated by forcing the state of the FSM to zero at the instant k D 4. For a general representation of the trellis, see Figure 12.13. Each path of the trellis corresponds to only one message sequence fb` g and is associated with only one sequence of code symbols fak g. The optimum decoder searches the trellis for the most probable path, given the received sequence fz k g is observed at the output of the demodulator. This search is usually realized by the Viterbi algorithm (see Section 8.10). Because of the presence of noise, the chosen path may not coincide with the correct one, but diverge from it at the instant k D i and rejoin it at the instant k D i C L; in this case we say that an error event of length L has occurred, as illustrated in the example in Figure 12.4 for an error event of length two (see Denition 8.1 on page 683). Note that in a trellis diagram more branches may connect the same pair of nodes. In this case we speak of parallel transitions, and by the term free distance of the code
972
k= 0 s0 = 0 1 2 3 4 5 6 7 0 4 2 6
40 6 2 5 1 7 3 1 5 3 7
Figure 12.3. Trellis diagram for the encoder of Figure 12.2 and the map of Figure 12.5. Each branch is labeled with the corresponding value of ak .
ak = sk = 0 1 2 3 4 5 6 7 ^ = ak
Figure 12.4. Section of the trellis for the decoder of an eight-state trellis code. The two continuous lines indicate two possible paths relative to two 8-PSK signal sequences, fak g and fak g.
we denote the minimum among the distances between symbols on parallel transitions and the distances between code sequences associated with pairs of paths in the trellis that originate from a common node and merge into a common node after L transitions, L > 1. By utilizing the sequence of samples fz k g, the decoding of a TCM signal is done in two phases. In the rst phase, called subset decoding, within each subset of symbols assigned
973
to the parallel transitions in the trellis diagram, the receiver determines the symbol closest to the received sample; these symbols are then memorized together with their squared distances from the received sample. In the second phase we apply the Viterbi algorithm to nd the code sequence fak g along the trellis such that the sum of the squared distances O between the code sequence and the sequence fz k g is minimum. Recalling that the signal is obtained at the output of the demodulator in the presence of additive white Gaussian noise with variance I2 per dimension, the probability of an error event for large values of the signal-to-noise ratio is approximated by (see (8.195)) dfree (12.4) Pe ' Nfree Q 2 I where dfree is dened in (12.2). From the Denition 6.2 on page 508 and the relation (12.4) between Euclidean distance and error probability, we give the denition of asymptotic coding gain, G code ;2 as the ratio between the minimum distance, dfree , between code sequences and the minimum Euclidean distance for uncoded sequences, equal to the minimum distance between symbols of the Q constellation of an uncoded system, 10 , normalized by the ratio between the average energy of the coded sequence, E s;c , and the average energy of the uncoded sequence, E s;u . The coding gain is then expressed in dB as G code D 10 log10
2 Q dfree =12 0
E s;c =E s;u
(12.5)
12.1.2
Set partitioning
The design of trellis codes is based on a method called mapping by set partitioning. This method requires that the bit mapper assign symbol values to the input binary vectors so that the minimum Euclidean distance between possible code sequences fak g is maximum. For a given encoder the search of the optimum assignment is made by taking into consideration subsets of the symbol set A. These subsets are obtained by successive partitioning of the set A, and are characterized by the property that the minimum Euclidean distance between symbols in a subset corresponding to a certain level of partitioning is larger than or equal to the minimum distance obtained at the previous level. Consider the symbol alphabet A D A0 with 2n elements, that corresponds to level zero of partitioning. At the rst level of partitioning, that is characterized by the index q D 1, the set A0 is subdivided into two disjoint subsets A1 .0/ and A1 .1/ with 2n 1 elements each. Let 11 .0/ and 11 .1/ be the minimum Euclidean distances between elements of the subsets A1 .0/ and A1 .1/, respectively; dene 11 as the minimum between the two Euclidean distances 11 .0/ and 11 .1/; we choose a partition for which 11 is maximum. At the level of partitioning characterized by the index q > 1, each of the 2q 1 subsets Aq 1 .`/, ` D 0; 1; : : : ; 2q 1 1, is subdivided into two subsets, thus originating 2q subsets. During
2
To emphasize the dependence of the asymptotic coding gain on the choice of the symbol constellations of the coded and uncoded systems, sometimes the information on the considered modulation schemes is included as a subscript in the symbol used to denote the coding gain, e.g. G 8PSK/4PSK for the introductory example.
974
the procedure, it is required that the minimum Euclidean distance at the q-th level of partitioning, 1q D
`2f0;1;:::;2q 1g
min
1q .`/
with
1q .`/ D
i ; m 2 Aq .`/ i 6D m
min
ji
m j
(12.6)
is maximum. At the n-th level of partitioning the subsets An .`/ consist of only one element each; to subsets with only one element we assign the minimum distance 1n D 1; at the end of the procedure we obtain a tree diagram of binary partitioning for the symbol set. At the q-th level of partitioning, to the two subsets obtained by a subset at the .q 1/-th level we assign the binary symbols y .q 1/ D 0 and y .q 1/ D 1, respectively; in this manner, an n-tuple of binary symbols yi D .yi.n 1/ ; : : : ; yi.1/ ; yi.0/ / is associated with each element i found at an end node of the tree diagram.3 Therefore the Euclidean distance between two elements of A, i and m , indicated by the binary vectors yi and ym that are equal in the rst q components, satises the relation ji m j 1q for yi
. p/
D ym
. p/
p D 0; : : : ; q
i 6D m
(12.7)
In fact, because of the equality of the components in the positions from .0/ up to .q 1/, we have that the two elements are in the same subset Aq .`/ at the q-th level of partitioning. Therefore their Euclidean distance is at least equal to 1q . Example 12.1.2 The partitioning of the set A0 of symbols with statistical power E[jak j2 ] D 1 for an 8-PSK system is illustrated in Figure 12.5. The minimum Euclidean distance between elements of the set A0 is given by 10 D 2 sin.=8/ D 0:765. At the rst level of partitioning the two subsets B0 D f.y .2/ ; y .1/ ; 0/; y .i / D 0; 1g and B1 D f.y .2/ ; y .1/ ; 1/; y .i / D 0; 1g are found, p with four elements each and minimum Euclidean distance 11 D 2. At the second level of partitioning four subsets C0 D f.y .2/ ; 0; 0/; y .2/ D 0; 1g, C2 D f.y .2/ ; 1; 0/; y .2/ D 0; 1g, C1 D f.y .2/ ; 0; 1/; y .2/ D 0; 1g, and C3 D f.y .2/ ; 1; 1/; y .2/ D 0; 1g are found with two elements each and minimum Euclidean distance 12 D 2. Finally, at the last level eight subsets D0 ; : : : ; D7 are found, with one element each and minimum Euclidean distance 13 D 1. Example 12.1.3 The partitioning of the set A0 of symbols with statistical power E[jak j2 ] D 1 for a 16-QAM system is illustrated in Figure 12.6. The minimum Euclidean distance between the elements p of A0 is given by 10 D 2= 10 D 0:632. Note that at each successive partitioning level the minimum Euclidean distance among the elements of a subset increases by a factor equal to p 2. Therefore at the third level of partitioning the minimum Euclidean distance between p the elements of each of the subsets Di , i D 0; 1; : : : ; 7, is given by 13 D 810 .
For TCM encoders, the n-tuples of binary code symbols will be indicated by y D .y .n 1/ ; : : : ; y .0/ / rather than by the notation c employed in the previous chapter.
975
Figure 12.5. Partitioning of the symbol set for an 8-PSK system. [From Ungerboeck (1982). c 1982 IEEE.]
Figure 12.6. Partitioning of the symbol set for a 16-QAM system. [From Ungerboeck (1982). c 1982 IEEE.]
12.1.3
Lattices
Several constellations and the relative partitioning can be effectively described by lattices; furthermore, as we will see in the following sections, the formulation based on lattices is particularly convenient in the discussion on multidimensional trellis codes.
976
In general, let Z D D Z D , where Z denotes the set of integers;4 a lattice 3 in < D is dened by the relation 3 D f.i 1 ; : : : ; i D / G j .i 1 ; : : : ; i D / 2 Z D g (12.8)
where G is a non-singular D D matrix, called lattice generator matrix, by means of which we obtain a correspondence Z D ! 3. The vectors given by the rows of G form a basis for the lattice 3; the vectors of the basis dene a parallelepiped whose volume V0 D j det.G/j represents the characteristic volume of the lattice. The volume V0 is equivalent to the volume of a Voronoi cell associated with an element or point of lattice 3 and dened as the set of points in < D whose distance from a given point of 3 is smaller than the distance from any other point of 3. The set of Voronoi cells associated with the points of 3 is equivalent to the space < D . A lattice is characterized by two parameters: 1. dmin , dened as the minimum distance between points of the lattice; 2. the kissing number, dened as the number of lattice points at minimum distance from a given point. We obtain a subgroup 3q .0/ if points of the lattice 3 are chosen as basis vectors in a matrix Gq , such that they give rise to a characteristic volume Vq D j det.Gq /j > V0 . Example 12.1.4 (Z p lattice) In general, as already mentioned, the notation Z p is used to dene a lattice with an innite number of points in the p-dimensional Euclidean space with coordinates given by integers. The generator matrix G for the lattice Z p is the p p identity matrix; the minimum distance is dmin D 1 and the kissing number is equal to 2 p. The Z2 type constellations (see Figure 12.7a) for QAM systems are nite subsets of Z2 , with center at the origin and minimum Euclidean distance equal to 10 . Example 12.1.5 (Dn lattice) Dn is the set of all n-dimensional points whose coordinates are integers that sum to an even number; it may be regarded as a version of the Zn lattice from which the points whose coordinates are integers that sum to an odd number were removed. The minimum distance is p dmin D 2 and the kissing number is 2n.n 1/. The lattice D2 is represented in Figure 12.7b. a D4 , called the Schl i lattice, constitutes the densest lattice in <4 ; this means that if four-dimensional spheres with centers given by the points of the lattice are used to ll <4 , then D4 is the lattice having the largest number of spheres per unit of volume. Example 12.1.6 (E8 lattice) E8 is given by points (
1 .x1 ; : : : ; x 8 / j 8i : xi 2 Z or 8i : xi 2 Z C ; 2
8 X i D1
) xi D 0 mod 2 (12.9)
977
3 2 1 0 1 1 2 3 4
(a)
2 1 0 1 2 3 4
(b)
Figure 12.7. (a) Z2 lattice; (b) D2 lattice.
In other words E8 is the set of eight-dimensional points whose components are all integers, or all halves of odd integers, that sum to an even number. E8 is called the Gosset lattice. We now discuss set partitioning with the aid of lattices. First we recall the properties of subsets obtained by partitioning. If the set A has a group structure with respect to a certain operation (see page 844), the partitioning can be done so that the sequence of subsets A0 ; A1 .0/; : : : ; An .0/, with Aq .0/ Aq 1 .0/, form a chain of subgroups of A0 ; in this case the subsets Aq .`/, ` 2 f1; : : : ; 2q 1g, are called cosets of the subgroup Aq .0/ with respect to A0 (see page 837), and are obtained from the subgroup Aq .0/ by translations. The distribution of Euclidean distances between elements of a coset Aq .`/ is equal to the distribution of the Euclidean distances between elements of the subgroup Aq .0/, as the difference between two elements of a coset yields an element of the subgroup; in particular, for the minimum Euclidean distance in subsets at a certain level of partitioning it holds 1q .`/ D 1q 8` 2 f0; : : : ; 2q 1g (12.10)
The lattice 3 in < D dened as in (12.8) has group structure with respect to the addition. With a suitable translation and normalization, we obtain that the set A0 for PAM or QAM is represented by a subset of Z or Z2 . To get a QAM constellation from Z2 , we dene, for example, the translated and normalized lattice Q D c.Z2 C f1=2; 1=2g/, where c is an arbitrary scaling factor, generally chosen to normalize the statistical power of the symbols to 1. Figure 12.8 illustrates how QAM constellations are obtained from Z2 . If we apply binary partitioning to the set Z or Z2 , we still get innite lattices in < or <2 , in which the minimum Euclidean distance increases with respect to the original lattice. Formally, we can assign the binary representations of the tree diagram obtained by
978
u u u u u u u u u u u u u u u u
u u u u u u u u u u u u u u u u
u u u u u u u u u u u u u u u u
u u u u u u u u u u u u u u u u
u u u u u u u u u u u u u u u u
u u u u u u u u u u u u u u u u
u u u u u u u u u u u u u u u u
u u u u u u u u u u u u u u u u
u u u u u u u u u u u u u u u u
u u u u u u u u u u u u u u u u
u u u u u u u u u u u u u u u u
u u u u u u u u u u u u u u u u
u u u u u u u u u u u u u u u u
partitioning to the lattices; for transmission, a symbol is chosen as representative for each lattice at an end node of the tree diagram. Denition 12.1 The notation X=X0 denotes the set of subsets obtained from the decomposition of the group X in the subgroup X0 and its cosets. The set X=X0 forms in turn a group, called the quotient group of X with respect to X0 . It is called binary if the number of elements is a power of two. PAM. In this case the subgroups of the lattice Z are expressed by Aq .0/ D f2q i j i 2 Zg D 2q Z Let t .`/ D ` 2 f0; : : : ; 2q 1g (12.12) (12.11)
979
then the cosets of a subgroup, expressed as Aq .`/ D fa C t .`/ j a 2 Aq .0/g are obtained by translations of the subgroup. The sequence Z = 2Z = 22 Z = : : : = 2q Z = : : : (12.14) (12.13)
forms a binary partitioning chain of the lattice Z, with increasing minimum distances given by 1q D 2q (12.15)
QAM. The rst subgroup in the binary partitioning chain of the two-dimensional lattice p Z2 is a lattice that is obtained from Z2 by rotation of =4 and multiplication by 2. The matrix of this linear transformation is 1 1 RD (12.16) 1 1 Successive subgroups in the binary partitioning chain are obtained by repeated application of the linear transformation R, Aq .0/ D Z2 Rq Let iZ2 Rq D f.ik; im/Rq j k; m 2 Zg where N is the set of natural numbers, then the sequence Z2 = Z2 R = Z2 R2 = Z2 R3 = D Z2 = Z2 R = 2Z2 = 2Z2 R = : : : (12.19) i 2N (12.18) (12.17)
forms a binary partitioning chain of the lattice Z2 , with increasing minimum distances given by 1q D 2q=2 (12.20)
This binary partitioning chain is illustrated in Figure 12.9. The cosets Aq .`/ are obtained by translations of the subgroup Aq .0/ as in the onedimensional case (12.13), with t.`/ D .i; m/ (12.21)
where, as it can be observed in Figure 12.10 for the case q D 2 with A2 .0/ D 2Z2 , q q q 2 2 i 2 0; : : : ; 2 1 m 2 0; : : : ; 2 1 ` D 22 i C m q even qC1 i 2 0; : : : ; 2 2 1 q 1 m 2 0; : : : ; 2 2 1 `D2
q 1 2 i
C m q odd (12.22)
980
Z2
Z2 R
2 Z2
2 Z2 R
2 Z2
2 Z 2 +(0,1)
2 Z 2 +(1,1)
2 Z 2 +(1,0)
M-PSK. In this case the set of symbols A0 D fe j2.k=M/ j k 2 Zg on the unit circle of the complex plane forms a group with respect to the multiplication. If the number of elements M is a power of two, the sequence o n 2q (12.23) A0 = A1 .0/ = A2 .0/ = : : : = Alog2 M .0/ with Aq .0/ D e j2 M k k 2 Z forms a binary partitioning chain of the set A0 , with increasing minimum distances 8 q 2 < for 0 q < log2 M 2 sin (12.24) 1q D M : 1 for q D log2 M The cosets of the subgroups Aq .0/ are given by n o ` Aq .`/ D a e j2 M a 2 Aq .0/
` 2 f0; : : : ; 2q
1g
(12.25)
12.1.4
As discussed in the previous subsections, an encoder can be modeled as a nite-state machine with a given number of states and well-dened state transitions. If the encoder input consists of m binary symbols per modulation interval,5 then there are 2m possible transitions from a state to the next state; there may be parallel transitions between pairs of states; furthermore for reasons of symmetry we take into consideration only encoders with a uniform structure. After selecting a diagram having the desired characteristics in terms
In this chapter encoding and bit mapping are jointly optimized and m represents the number of information bits L b per modulation interval (see (6.93)). For example, for QAM the rate of the encoder-modulator is R I D m . 2
981
of state transitions, the design of a code is completed by assigning symbols to the state transitions, such that dfree is maximum. Following the indications of information theory (see Section 6.10), the symbols are chosen from a redundant set A of 2mC1 elements. Example 12.1.7 (Uncoded transmission of 2 bits per modulation interval by 8-PSK) Consider an uncoded 4-PSK system as reference system. The uncoded transmission of 2 bits per modulation interval by 4-PSK can be viewed as the result of the application of a trivial encoder with only one state, and trellis diagram with four parallel transitions, as illustrated in Figure 12.11. A distinct symbol of the alphabet of a 4-PSK system is assigned to each parallel transition. Note from Figure 12.5 that the alphabet of a 4-PSK system is obtained by partitioning the alphabet A0 of an obtained choosing the subset B0 (or B1 )p 8-PSK system. Therefore dmin D 11 D 2. In the trellis diagram, any sequence on the trellis represents a possible symbol sequence. The optimum receiver decides for the symbol of the subset B0 (or B1 ) that is found at the minimum distance from the received signal. Now consider the two-state trellis diagram of Figure 12.12. The symbols of the subsets B0 D C0 [ C2 and B1 D C1 [ C3 are assigned to the transitions that originate from the rst and second state, respectively; this guarantees that the minimum dfree between the code symbol sequences is at least equal to that obtained for uncoded 4-PSK system. With a trellis diagram with only two states it is impossible to have only signals of B0 or B1 assigned to the transitions that originate from a certain state, and also to all transitions that lead to the same state; therefore we nd that the minimum Euclidean distance between q code symbol sequences is in this case equal to dfree D 12 C 12 D 1:608, greater than 1 0 Q that obtained for uncoded 4-PSK transmission. As E s;c D E s;u and 10 D 11 , the coding q q 2 C 12 = 12 / D 1:1 dB. Furthermore N gain is G 8PSK/4PSK D 20 log10 . 11 free D 2. 0 1 High coding gains for the transmission of 2 bits per modulation interval by 8-PSK are obtained by the codes represented in Figure 12.13, with trellis diagrams having 4, 8, and 16 states. For the heuristic design of these codes with moderate complexity we resort to the following rules proposed by Ungerboeck: 1. all symbols of the set A0 must be assigned equally likely to the state transitions in the trellis diagram, using criteria of regularity and symmetry;
982
Figure 12.12. Transmission of 2 bits per modulation interval using a two-state trellis code and 8-PSK. For each state the values of the symbols assigned to the transitions that originate from that state are indicated.
Figure 12.13. Trellis codes with 4, 8, and 16 states for transmission of 2 bits per modulation interval by 8-PSK. [From Ungerboeck (1982). c 1982 IEEE.]
983
2. to transitions that originate from the same state are assigned symbols of the subset B0 , or symbols of the subset B1 ; 3. to transitions that merge to the same state are assigned symbols of the subset B0 , or symbols of the subset B1 ; 4. to parallel transitions between two states we assign the symbols of one of the subsets C0 , C1 , C2 , or C3 . Rule 1 intuitively points to the fact that good trellis codes exhibit a regular structure. Rules 2, 3, and 4 guarantee that the minimum Euclidean distance between code symbol sequences that differ in one or more elements is at least twice the minimum Euclidean distance between uncoded 4-PSK symbols, so that the coding gain is greater than or equal to 3 dB, as we will see in the next examples. Example 12.1.8 (Four-state trellis code for the transmission of 2 bit/s/Hz by 8-PSK) Consider the code with four states represented in Figure 12.13. Between each pair of code symbol sequences in the trellis diagram that diverge at a certain state and q merge after more than one transition, the Euclidean distance is greater than or equal to 12 C 12 C 12 1 0 1 D 2:141. For example, this distance exists between sequences in the trellis diagram labeled by the symbols 000 and 212; on the other hand, the Euclidean distance between symbols assigned to parallel transitions is equal to 12 D 2. Therefore the minimum Euclidean distance between code symbol sequences is equal to 2; hence with a four-state trellis code we obtain a gain equal to 3 dB over uncoded 4-PSK transmission. Note that, as the minimum distance dfree is determined by parallel transitions, the sequence at minimum distance from a transmitted sequence differs only by one element that corresponds to the transmitted symbol rotated by 180 . A possible implementation of the encoder/bit-mapper for a four-state trellis code is illustrated in Figure 12.14. The 4-state trellis code for transmission with spectral efciency of 2 bit/s/Hz by 8-PSK was described in detail as an introductory example. In Figure 12.13 the values of dfree and Nfree for codes with 4, 8, and 16 states are reported.
Figure 12.14. Encoder/bit-mapper for a 4-state trellis code for the transmission of 2 bits per modulation interval by 8-PSK.
984
Figure 12.15. Eight-state trellis code for the transmission of 3 bits per modulation interval by 16-QAM.
Consider now a 16-QAM system for the transmission of 3 bits per modulation interval; in this case the reference system uses uncoded 8-PSK or 8-AM-PM, as illustrated in Figure 12.15. Example 12.1.9 (Eight-state trellis code for the transmission of 3 bit/s/Hz by 16-QAM) The partitioning of a symbol set with unit statistical power for a 16-QAM system is shown in Figure 12.6. For the assignment of symbols to the transitions on the trellis consider the elements each. subsets of the symbol set A0 denoted by D0 ; D1 ; : : : ; D7 , that contain two p The minimum Euclidean distance between the signals in A0 is 10 D 2= 10 D 0:632; the minimum Euclidean distance between elements of a subset Di , i D 0; 1; : : : ; 7, is p 13 D 810 . In the 8-state trellis code illustrated in Figure 12.15, four transitions diverge from each state and four merge to each state. To each transition one of the subsets Di , i D 0; 1; : : : ; 7, is assigned; therefore a transition in the trellis corresponds to a pair of parallel transitions. The assignment of subsets to the transitions satises Ungerboeck rules. The subsets D0 , D4 , D2 , D6 , or D1 , D5 , D3 , D7 are assigned to the four transitions from or to the same state. In evaluating dfree , this choice guarantees a squared Euclidean distance equal to at least 212 between sequences that diverge from a state and merge after L transitions, L > 1. The 0 squared distance between sequences that diverge from a state and merge after two transitions is equal to 612 . If two sequences diverge and merge again after three or more transitions, 0 at least one intermediate transition contributes to an incremental squared Euclidean distance equal to 12 ; thus the minimum Euclidean distance between code symbol sequences that 0 p do not differ only for one symbol is given by 510 . As the Euclidean distance between p symbols assigned to parallel transitions is equal to 810 , the free distance of the code p is dfree D 510 . Because the minimum Euclidean distance for an uncoded 8-AM-PM p Q reference system with the same average symbol energy is 10 D 210 , the coding gain is p G 16QAM/8AM-PM D 20 log10 f 5=2g D 4 dB.
985
In the trellis diagram of Figure 12.15 four paths are shown that represent error events at minimum distance from the code sequence, having symbols taken from the subsets D0 D0 D3 D6 ; the sequences in error diverge from the same state and merge after three or four transitions. It can be shown that for each code sequence and for each state there are two paths leading to error events of length three and two of length four. The number of code sequences at the minimum distance depends on the code sequence being considered, hence Nfree represents a mean value. The proof is simple in the case of uncoded 16-QAM, where the number of symbols at the minimum distance from a symbol that is found at the center of the constellation is larger than the number of symbols at the minimum distance from a symbol found at the edge of the constellation; in this case we obtain Nfree D 3. For the eight-state trellis code of Figure 12.15, we get Nfree D 3:75. For constellations of type Z2 with a number of signals that tends to innity, the limit is Nfree D 4 for uncoded modulation, and Nfree D 16 for coded modulation with an eight-state code.
12.1.5
The structure of trellis codes for TCM can be described in general by the combination of a convolutional encoder and a special bit mapper, as illustrated in Figure 12.16. A code .m/ .1/ symbol sequence is generated as follows. Of the m information bits [bk ; : : : ; bk ] that must be transmitted during a cycle of the encoder/bit-mapper operations, m m are input Q Q Q Q to a convolutional encoder with rate Rc D m=.m C 1/; the m C 1 bits at the encoder output, .Q .0/ Q Q yk D [ykm/ ; : : : ; yk ]T , are used to select one of the 2mC1 subsets of the symbol set A with mC1 elements, and to determine the next state of the encoder. The remaining .m m/ uncoded 2 Q Q bits determine which of the 2m m symbols of the selected subset must be transmitted. For example, in the encoder/bit-mapper for a four-state code illustrated in Figure 12.14, the two .0/ .1/ bits yk and yk select one of the four subsets C0 ; C2 ; C1 ; C3 of the set A0 with 8 elements. .2/ The uncoded bit yk determines which of the two symbols in the subset Ci is transmitted.
986
.m/ .1/ .0/ Let yk D [yk ; : : : ; yk ; yk ] be the .m C 1/-dimensional binary vector at the input of the bit-mapper at the k-th instant, then the selected symbol is expressed as ak D a[yk ]. Note Q that in the trellis the symbols of each subset are associated with 2m m parallel transitions. The free Euclidean distance of a trellis code, given by (12.2), can be expressed as
(12.26)
where 1mC1 is the minimum distance between symbols assigned to parallel transitions and Q Q dfree .m/ denotes the minimum distance between code sequences that differ in more than one symbol. In the particular case m D m, each subset has only one element and therefore Q there are no parallel transitions; this occurs, for example, for the encoder/ bit-mapper for an 8-state code illustrated in Figure 12.2. From Figure 12.16, observe that the vector sequence fQ k g is the output sequence of a y convolutional encoder. Recalling (11.263), for a convolutional code with rate m=.m C 1/ Q Q and constraint length , we have the following constraints on the bits of the sequence fQ k g: y
m Q X i D0 .i .i .i h .i / yk / h .i / 1 yk / C1 h .i / yk / D 0 0
8k
(12.27)
where fh .i / g, 0 j , 0 i m, are the parity check binary coefcients of the encoder. Q j For an encoder having binary memory cells, a trellis diagram is generated with 2 states. Note that (12.27) denes only the constraints on the code bits, but not the input/output relation of the encoder. Using polynomial notation, for the binary vector sequence y.D/ (12.27) becomes [y .m/ .D/; : : : ; y .1/ .D/; y .0/ .D/] [h .m/ .D/; : : : ; h .1/ .D/; h .0/ .D/]T D 0 (12.28)
Q where h .i / .D/ D h .i / D C h .i / 1 D 1 C C h .i / D C h .i / , for i D 0; 1; : : : ; m, and 1 0 Q h .i / .D/ D 0 for m < i m. From (12.28) we observe that the code sequences y.D/ can be obtained by a systematic encoder with feedback as6 3T 2 0 2 .m/ 3 6 : 7 2 3 : y .D/ 7 6 : b.m/ .D/ 7 6 6 7 6 : 7 6 : 0 7 6 7 6 : : 7 4 : (12.29) 6 7 D 6 Im 5 Q : .1/ .D/ 5 h .m/ .D/= h .0/ .D/ 7 4 y 7 6 .1/ .D/ 7 6 b : : y .0/ .D/ 5 4 : h .1/ .D/= h .0/ .D/ The rational functions h .i / .D/= h .0/ .D/, i D 1; : : : ; m, are realizable if the following conQ dition is satised [9]: ( ) 0 i 6D 0 .i / .i / h0 D h D 2 (12.30) 1 i D0
Note that (12.29) is analogous to (11.304); here the parity check coefcients are used.
987
b (m) k
(m b k +1) b (m) k
y (1) k
y (0) k
Figure 12.17. Block diagram of a systematic convolutional encoder with feedback. [From Ungerboeck (1982). c 1982 IEEE.]
The implementation of a systematic encoder with feedback having binary memory elements is illustrated in Figure 12.17.
Computation of dfree
Consider the two code sequences
.m/ .0/ y1 .D/ D [y1 .D/; : : : ; y1 .D/]T
(12.31) (12.32)
.m/ .0/
.m/
.0/
related by y2 .D/ D y1 .D/ e.D/, as the code is linear. Let ek D [ek ; : : : ; ek ], then the error sequence e.D/ is given by e.D/ D ek D k C ekC1 D kC1 C C ekCL D kCL (12.33)
where ei ; ei CL 6D 0, and L > 0. Note that e.D/ is also a valid code sequence as the code is linear. To nd a lower bound on the Euclidean distance between the code symbol sequences a1 .D/ D a[y1 .D/] and a2 .D/ D a[y2 .D/] obtained from y1 .D/ and y2 .D/, we dene the function d[ek ] D minzk d.a[zk ]; a[zk ek ]/, where minimization takes place in the space .m/ .1/ .0/ of the binary vectors zk D [z k ; : : : ; z k ; z k ]T , and d.;/ denotes the Euclidean distance between specied symbols. For the squared distance between the sequences a1 .D/ and a2 .D/ then the relation holds
i CL X kDi
d 2 .a[yk ]; a[yk ek ]/
i CL X kDi
d 2 [ek ] D d 2 [e.D/]
(12.34)
988
Theorem 12.1 For each sequence e.D/ there exists a pair of symbol sequences a1 .D/ and a2 .D/ for which relation (12.34) is satised with the equal sign. Proof. Due to the symmetry in the subsets of symbols obtained by partitioning, d[ek ] D .0/ minzk d.a[zk ]; a[zk ek ]/ can arbitrarily be obtained by letting the component z k of the vector zk equal to 0 or to 1 and performing the minimization only with respect to .m/ .1/ components [z k ; : : : ; z k ]. As encoding does not impose any constraint on the component .m/ .1/ sequence [yk ; : : : ; yk ],7 for each sequence e.D/ a code sequence y.D/ exists such that the relation (12.34) is satised as equality for every value of the index k. The free Euclidean distance between code symbol sequences can therefore be determined by a method similar to that used to nd the free Hamming distance between binary code sequences y.D/ (see (11.266)). We need to nd an efcient algorithm to examine all possible error sequences e.D/ (12.33) and to substitute the squared Euclidean distance d 2 [ek ] to the Hamming weight of ek ; thus
2 dfree .m/ D Q e.D/Dy2 .D/ y1 .D/6D0
min
i CL X kDi
d 2 [ek ]
(12.35)
Let q.ek / be the number of consecutive components equal to zero in the vector ek , .0/ .m/ .3/ starting with component ek . For example, if ek D [ek ; : : : ; ek ; 1; 0; 0]T , then q.ek / D 2. From the denition of the indices assigned to symbols by partitioning, we obtain that d[ek ] 1q.ek / ; moreover this relation is satised as equality for almost all vectors ek . Note that d[0] D 1q.0/ D 0. Therefore
2 dfree .m/ min Q i CL X kDi 2 1q.ek / D 12 .m/ free Q
e.D/6D0
(12.36)
Q Q If we assume dfree .m/ D 1free .m/, the risk of committing an error in evaluating the free distance of the code is low, as the minimum is usually reached by more than one error sequence. By the denition of free distance in terms of the minimum distances between Q elements of the subsets of the symbol set, the computation of dfree .m/ will be independent of the particular assignment of the symbols to the binary vectors with .m C 1/ components, provided that the values of the minimum distances among elements of the subsets are not changed. At this point it is possible to identify a further important consequence of the constraint (12.30) on the binary coefcients of the systematic convolutional encoder. We can show that an error sequence e.D/ begins with ei D .ei.m/ ; : : : ; ei.1/ ; 0/ and ends with ei CL D .ei.m/ ; : : : ; ei.1/ ; 0/. It is therefore guaranteed that all transitions that originate from the CL CL same state and to the transitions that merge at the same state are assigned signals of the subset B0 or those of the subset B1 . The squared Euclidean distance associated with an error sequence is therefore greater than or equal to 212 . The constraint on the parity check 1
7
From the parity equation (12.29) we observe that a code sequence fzk g can have arbitrary values for each .m/ .1/ m-tuple [z k ; : : : ; z k ].
989
coefcients allows us, however, to determine only a lower bound for dfree .m/. For a given Q sequence 10 11 1mC1 of minimum distances between elements of subsets Q and a code with constraint length , a convolutional code that yields the maximum value Q of dfree .m/ is usually found by a computer program for code search. The search of the . 1/.m C 1/ parity check binary coefcients is performed by means such that the explicit Q Q computation of dfree .m/ is often avoided. Tables 12.1 and 12.2 report the optimum codes for TCM with symbols of the type Z1 and Z2 , respectively [2]. For 8-PSK, the optimum codes are given in Table 12.3 [2].
Table 12.1 Codes for one-dimensional modulation. [From Ungerboeck (1987). c 1987 IEEE.]
2 4 8 16 32 64 128
m Q
h1
4 4 8 12 36 66
Table 12.2 Codes for two-dimensional modulation. [From Ungerboeck (1987). c 1987 IEEE.]
2 m Q h2 h1 h0
2 Q dfree =12 0
Nfree .m ! 1/ 4 16 56 16 56 344 44 4
1 2 2 2 2 2 2 2
Table 12.3 Codes for 8-PSK. [From Ungerboeck (1987). c 1987 IEEE.]
2 4 8 16 32 64 128
m Q 1 2 2 2 2 2
h2 04 16 34 066 122
h1 2 02 04 16 030 054
h0 5 11 23 45 103 277
2 Q dfree =12 0
990
Parity check coefcients are specied in octal notation; for example, the binary vector .0/ .0/ [h 6 ; : : : ; h 0 ] D [1; 0; 0; 0; 1; 0; 1] is represented by h.0/ D 1058 . In the tables, an asterisk next to the value dfree indicates that the free distance is determined by the parallel transitions, that is dfree .m/ > 1mC1 . Q Q
12.2
Multidimensional TCM
So far we have dealt with trellis coded modulation schemes that use two-dimensional (2D) constellations, that is to send m information bits per modulation interval we employ a 2D constellation of 2.mC1/ points; the intrinsic cost is represented by doubling the cardinality of the 2D constellation with respect to uncoded schemes, as a bit of redundancy is generated at each modulation interval. Consequently the minimum distance within points of the constellation is reduced, for the same average power of the transmitted signal; without this cost the coding gain would be 3 dB higher. An advantage of using a multidimensional constellation to generate code symbol sequences is that doubling the cardinality of a constellation does not lead to a 3 dB loss; in other words, the signals are spaced by a larger Euclidean distance dmin and therefore the margin against noise is increased. A simple way to generate a multidimensional constellation is obtained by time division. If, for example, ` two-dimensional symbols are transmitted over a time interval of duration Ts , and each of them has a duration Ts =`, we may regard the ` 2D symbols as an element of a 2`-dimensional constellation. Therefore in practice multidimensional signals can be transmitted as sequences of one or two-dimensional symbols. An example of multidimensional signaling using binary PAM transmission is illustrated in Section 6.8. In this section we describe the construction of 2`-dimensional TCM schemes for the transmission of m bits per 2D symbol, and thus m` bits per 2`-dimensional signal. We maintain the principle of using a redundant signal set, with a number of elements doubled with respect to that used for uncoded modulation; therefore the 2`-dimensional TCM schemes use sets of 2m`C1 2`-dimensional signals. With respect to two-dimensional TCM schemes, this implies a lower redundancy in the two-dimensional component sets. For example, in the 4D case doubling the p number of elements causes an expansion of the 2D component constellations by a factor 2; this corresponds to a half bit of redundancy per 2D component constellation. The cost of the expansion of the signal set is reduced by 1.5 dB in the 4D case and by 0.75 dB in the 8D case. A further advantage of multidimensional constellations is that the design of schemes invariant to phase rotation is simplied (see Section 12.3).
Encoding
Starting with a constellation A0 of the type Z1 or Z2 in the one or two-dimensional signal space, we consider multidimensional trellis codes where the signals to be assigned to the transitions in the trellis diagram come from a constellation A0 , I > 2, in the I multidimensional space < I . In practice, if .m C 1/ binary output symbols of the nite state encoder determine the assignment of modulation signals, it is possible to associate with these binary symbols a sequence of ` modulation signals, each in the constellation A0 , transmitted during ` modulation intervals, each of duration T ; this sequence can be
991
considered as an element of the space < I , as signals transmitted in different modulation intervals are assumed orthogonal. If we have a constellation A0 with M elements, the relation M ` 2mC1 holds. Possible sequences of ` symbols of a constellation A0 give origin to a block code B of length ` in the space < I . Hence, we can consider the multidimensional TCM as an encoding method in which the binary output symbols of the nite state encoder represent the information symbols of the block code. Part of the whole coding gain is obtained by choosing a multidimensional constellation A0 with a large minimum Euclidean distance, I that is equivalent to the choice of an adequate block code; therefore it is necessary to identify block codes in the Euclidean space < I that admit a partitioning of code sequences in subsets such that the minimum Euclidean distance among elements of a subset is the largest possible. For a linear block code, it can be shown that this partitioning yields as subsets a subgroup of the code and its cosets. Consider the trellis diagram of a code for Q the multidimensional TCM where each state has 2mC1 transitions to adjacent states, and m m parallel transitions. Then the construction Q between pairs of adjacent states there exist 2 Q of the trellis code requires that the constellation A0 is partitioned into 2mC1 block codes I j Q BmC1 , j 2 f0; : : : ; 2mC1 1g, each of length ` and rate equal to .m m/=I bits/dimension. Q Q In any period equivalent to ` modulation intervals, .m C 1/ binary output symbols of Q the nite state encoder select one of the code blocks, and the remaining .m m/ binary Q symbols determine the sequence to be transmitted among those belonging to the selected code. From (6.103), for a constellation A0 of type Z2 , Bmin Ts D `, and M ` D 2`mC1 , the spectral efciency of a system with multidimensional TCM is equal to D ` log2 M m D ` ` 1 D log2 M 1 ` bit/s/Hz (12.37)
The optimum constellation in the space < I is obtained by solving the problem of nding a lattice such that, given the minimum Euclidean distance dmin between two points, the number of points per unit of volume is maximum; for I D 2, the solution is given by the hexagonal lattice. For a number of constellation points M 1, the ratio between the statistical power of signals of a constellation of the type Z2 and that of signals of a constellation chosen as a subset of the hexagonal lattice is equal to 0.62 dB; hence, a constellation subset of the hexagonal lattice yields a coding gain equal to 0.62 dB with respect to a constellation of the type Z2 . For I D 4 and I D 8 the solutions are given by the Schl i lattice D4 and the Gosset a lattice E8 , respectively, dened in Examples 12.1.5 and 12.1.6 of page 976. Note that to an increase in the number of dimensions and in the density of the optimum lattice also corresponds an increase in the number of lattice points with minimum distance from a given point. To design codes with a set of modulation signals whose elements are represented by symbol sequences, we have to address the problem of partitioning a lattice in a multidimensional space. In the multidimensional TCM, if m C 1 binary output symbols of the nite state encoder Q Q determine the next state to a given state, the lattice 3 must be partitioned into 2mC1 subsets mC1 Q 3mC1 . j/, j D 0; : : : ; 2 1, such that the minimum Euclidean distance 1mC1 between Q Q the elements of each subset is maximum; hence, the problem consists in determining the Q subsets of 3 so that the density of the points of 3mC1 . j/, j D 0; : : : ; 2mC1 1, is maximum. Q
992
In the case of an I -dimensional lattice 3 that can be expressed as the Cartesian product of ` terms all equal to a lattice in the space < I =` , the partitioning of 3 can be derived from the partitioning of the I =`-dimensional lattice. Example 12.2.1 (Partitioning of the lattice Z4 ) The lattice A0 D Z4 can be expressed in terms of the Cartesian product of the lattice Z2 4 with itself, Z4 D Z2 Z2 ; the subsets of the lattice A0 are therefore characterized by two 4 sets of signals belonging to a Z2 lattice and by their subsets. The partitioning of a signal set belonging to a Z2 lattice is represented in Figure 12.6. Thus Z4 D A 0 D A 0 4 D .B0 A0 D .B0 [ B1 / B1 / [ .B1 .B0 [ B1 / B0 / [ .B1 B1 / (12.38)
B0 / [ .B0
B0 / [ .B1 B1 / [ .B1
B1 / B0 /
(12.39) (12.40)
In terms of the four-dimensional TCM, the assignment of pairs of two-dimensional modulation signals to the transitions in the trellis diagram is such that in the rst modulation interval all the points of the component QAM constellation are admissible. If two pairs assigned to adjacent transitions are such that the signals in the rst modulation interval are separated by the minimum Euclidean distance 10 , the signals in the second modulation interval will also have Euclidean distance at least equal to 10 ; in this way the minimum p 0 0 1 Euclidean distance among points of B4 or B4 is equal to 11 D 210 . The subset B4 is 1 the Schl i lattice D4 , that is the densest lattice in the space <4 ; the subset B4 is instead a the coset of D4 with respect to Z4 . The next partitioning in the four subsets B0 B0 B1 B1 B0 B1 B0 B1 (12.41)
does not yield any increase in the minimum Euclidean distance. Note that the four subsets at the second level of partitioning differ from the Z4 lattice only by the position with respect to the origin, direction, and scale; therefore the subsets at successive levels of partitioning are obtained by iterating the same procedure described for the rst two. Thus we have the following partitioning chain Z4 = D4 = .Z2 R/2 = 4D4 = : : : (12.42)
where R is the 2 2 matrix given by (12.16). The partitioning of the Z4 lattice is illustrated in Figure 12.18 [2]. Optimum codes for the multidimensional TCM are found in a similar manner to that described for the one- and two-dimensional TCM codes. Codes and relative asymptotic
993
Figure 12.18. Partitioning of the lattice A0 D Z4 . [From Ungerboeck (1987). c 1987 IEEE.] 4 Table 12.4 Codes for four-dimensional modulation. [From Ungerboeck (1987). c 1987 IEEE.]
2 8 16 32 64 128
m Q 2 2 3 4 4
h4 050 120
h3 30 030 050
h2 04 14 14 014 022
h1 02 02 02 002 006
h0 11 21 41 101 203
2 Q dfree =12 0
Nfree .m ! 1/ 88 24 8 144
coding gains for the four-dimensional TCM, with respect to uncoded modulation with signals of the type Z2 , are reported in Table 12.4 [2]. These gains are obtained for signal sets with a large number of elements that, in the signal space, take up the same volume of elements of the signal set used for uncoded modulation; then the comparison is made for the same statistical power and the same peak power of the two-dimensional signals utilized for uncoded modulation.
Decoding
The decoding of signals generated by multidimensional TCM is achieved by a sequence of operations that is the inverse with respect to the encoding procedure described in the previous subsection. The rst stage of decoding consists in determining, for each modulation interval, the Euclidean distance between the received sample and all M signals of the constellation A0 of symbols and also, within each subset Aq .i/ of the constellation A0 , the signal ak .i/ that has the minimum Euclidean distance from the received sample. The O Q second stage consists in the decoding, by a maximum likelihood decoder, of 2mC1 block 0 /. Due to the large number M ` of signals in the multidimensional space codes BmC1 .` Q < I , in general the block codes have a number of elements such that the complexity of a maximum likelihood decoder that should compute the metric for each element would result excessive. The task of the decoder can be greatly simplied thanks to the method followed
994
for the construction of block codes BmC1 .`0 /. Block codes are identied by the subsets Q of the multidimensional constellation A0 , which are expressed in terms of the Cartesian I product of subsets Aq .i/ of the one- or two-dimensional constellation A0 . Decoding of the block codes is jointly carried out by a trellis diagram dened on a nite number ` of modulation intervals, where Cartesian products of subsets of the constellation A0 in different modulation intervals are represented as sequences of branches, and the union of subsets as the union of branches. Figure 12.19 illustrates the trellis diagram for the decoding of block codes obtained by the partitioning into two, four, and eight subsets of a constellation A 0 Z4 . 4 As the decisions taken in different modulation intervals are independent, the branch O metrics of the trellis diagram is the squared Euclidean distance d 2 .z k ; ak .i// between the received sample during the k-th modulation interval and the signal ak .i/ of the subset Aq .i/ O at the minimum Euclidean distance from z k . The maximum likelihood decoder for block codes can then be implemented by the Viterbi algorithm, which is applied for ` iterations Q from the initial node of the trellis diagram to the 2mC1 terminal nodes; the procedure to decode block codes by a trellis diagram is due to Wolf [10]. The third and last decoding Q stage consists in using the metrics, obtained at the 2mC1 terminal nodes of the trellis diagram for the decoding of the block codes, in the Viterbi decoder for the trellis code, that yields O .m/ O .1/ the sequence of detected binary vectors .bk ; : : : ; bk /. To evaluate the complexity of each iteration in multidimensional TCM decoding, it is necessary to consider the overall number of branches in the trellis diagrams considered at the different decoding stages; this number includes the `M branches for decisions on signals within the subsets Aq .i/ in ` modulation intervals, the N R branches in the trellis Q diagram for the decoding of the block codes, and the 2Cm branches in the trellis diagram
B0 B0 B1 B1 B0 B1 4 C0 C2 C3 C1 B0 4 C0 C2 C0 4 C2 C0 C1 C3 C5 4 C2 4 C3 C1 C2 C0 B1 C2 C0 C7 4 C6 4 C3 4 C4 4 C1 4
B1
B0 B0 B1 B0
C1
C1 C3
C3
B1
Figure 12.19. Trellis diagram for the decoding of block codes obtained by the partitioning the lattice A0 D Z4 into two, four, and eight subsets. 4
995
for the decoding of the trellis code. For a code with efciency equal to .log2 M 1=`/ bits per modulation interval, the complexity expressed as the number of branches in the trellis diagram per transmitted information bit is thus given by
Q `M C N R C 2Cm ` log2 M 1
(12.43)
The number N R can be computed on the basis of the particular choice of the multidimensional constellation partitioning chain; for example, in the case of partitioning of the lattice Z4 into four or eight subsets, for the four-dimensional TCM we obtain N R D 20. Whereas in the case of two-dimensional TCM the decoding complexity essentially lies in the implementation of the Viterbi algorithm to decode the trellis code, in the four-dimensional TCM most of the complexity is due to the decoding of the block codes. In multidimensional TCM it is common to nd codes with a large number Nfree of error events at the minimum Euclidean distance; this characteristic is due to the fact that in dense multidimensional lattices the number of points at the minimum Euclidean distance rapidly increases with the increase in the number of dimensions. In this case the minimum Euclidean distance is not sufcient to completely characterize code performance, as the difference between asymptotic coding gain and effective coding gain, for values of interest of the error probability, cannot be neglected.
12.3
In PSK or QAM transmission schemes with coherent demodulation, an ambiguity occurs at the receiver whenever we ignore the value of the carrier phase used at the transmitter. In BPSK, for example, there are two equilibrium values of the carrier phase synchronization system at the receiver, corresponding to a rotation of the phase equal to 0 and 180 , respectively; if the rotation of the phase is equal to 180 , the binary symbols at the demodulator output come out inverted. In QAM, instead, because of the symmetry of the Z2 lattice points, there are four possible equilibrium values. The solution to the problem is represented by the insertion in the transmitted sequence of symbols for the synchronization, or by differential encoding (see Section 6.5.2). We recall that with differential encoding the information symbols are assigned to the phase difference between consecutive elements in the symbol sequence; in this way, the absolute value of phase in the signal space becomes irrelevant to the receiver. With TCM we obtain code signal sequences that, in general, do not present symmetries with respect to the phase rotation in the signal space. This means that, for a code symbol sequence, after a phase rotation determined by the phase synchronization system, we may obtain a sequence that does not belong to the code; therefore trellis coding not invariant to phase rotation can be seen as a method for the construction of signal sequences that allow the recovery of the absolute value of the carrier phase. Then, for the demodulation, we choose the value of the carrier phase corresponding to an equilibrium value of the synchronization system, for which the Euclidean distance between the received sequence and the code sequence is minimum. For fast carrier phase synchronization, various rotationally invariant trellis codes have been developed; these codes are characterized by the property that code signal sequences
996
continue to belong to the code even after the largest possible number of phase rotations. In this case, with differential encoding of the information symbols, the independence of demodulation and decoding from the carrier phase recovered by the synchronization system Q is guaranteed. A differential decoder is then applied to the sequence of binary symbols fb` g at the output of the decoder for the trellis code to obtain the desired detection of the sequence O of information symbols fb` g. Rotationally invariant trellis codes were initially proposed by Wei [11, 12]. In general, the invariance to phase rotation is more easily obtained with multidimensional TCM; in the case of two-dimensional TCM for PSK or QAM systems, it is necessary to use non-linear codes in GF(2). In the case of TCM for PSK systems, the invariance to phase rotation can be directly obtained using PSK signals with differential encoding. The elements of the symbol sets are not assigned to the binary vectors of the convolutional code but rather to the phase differences relative to the previous symbols.
Bibliography
[1] G. Ungerboeck, Channel coding with multilevel/phase signals, IEEE Trans. on Information Theory, vol. 28, pp. 5567, Jan. 1982. [2] G. Ungerboeck, Trellis coded modulation with redundant signal sets. Part I and Part II, IEEE Communications Magazine, vol. 25, pp. 621, Feb. 1987. [3] S. S. Pietrobon, R. H. Deng, A. Lafanechere, G. Ungerboeck, and D. J. Costello Jr., Trellis-coded multidimensional phase modulation, IEEE Trans. on Information Theory, vol. 36, pp. 6389, Jan. 1990. [4] E. Biglieri, D. Divsalar, P. J. McLane, and M. K. Simon, Introduction to trellis-coded modulation with applications. New York: Macmillan Publishing Company, 1991. [5] S. S. Pietrobon and D. J. Costello Jr., Trellis coding with multidimensional QAM signal sets, IEEE Trans. on Information Theory, vol. 39, pp. 325336, Mar. 1993. [6] J. Huber, Trelliscodierung. Heidelberg, Germany: Springer-Verlag, 1992. [7] C. Schlegel, Trellis coding. New York: IEEE Press, 1997. [8] G. D. Forney, Jr. and G. Ungerboeck, Modulation and coding for linear Gaussian channels, IEEE Trans. on Information Theory, vol. 44, pp. 23842415, Oct. 1998. [9] G. D. Forney, Jr., Convolutional codes I: algebraic structure, IEEE Trans. on Information Theory, vol. IT16, pp. 720738, Nov. 1970. [10] J. K. Wolf, Efcient maximum likelihood decoding of linear block codes using a trellis, IEEE Trans. on Information Theory, vol. IT24, pp. 7680, Jan. 1978.
12. Bibliography
997
[11] L. F. Wei, Trellis-coded modulation with multidimensional constellations, IEEE Trans. on Information Theory, vol. 33, pp. 483501, July 1987. [12] L. F. Wei, Rotationally invariant trellis-coded modulations with multidimensional M-PSK, IEEE Journal on Selected Areas in Communications, vol. 7, pp. 12811295, Dec. 1989.
Chapter 13
In this chapter we rst extend the study of the capacity of ideal AWGN channels introduced in Section 6.10 to the case of band limited dispersive channels with additive Gaussian noise. We nd that the optimum power spectral density of the transmitted signal is inversely proportional to the spectral signal-to-noise ratio, and is determined by a water pouring criterion. Then we discuss practical methods to approximate this capacity. In particular, we consider OFDM or multicarrier modulation described in Chapter 9, as well as single-carrier modulation in the form of CAP or QAM as described in Chapter 7, combined with joint precoding and coding techniques.
13.1
In general, a non-ideal channel (linear dispersive channel) with additive Gaussian noise is characterized as a transmission medium with impulse response gCh .t/ and additive Gaussian noise w.t/ having power spectral density P w . f /. Let s.t/ denote the channel input signal; then the receiver input signal is given by r.t/ D s gCh .t/Cw.t/. In practice, the transmitted signal s.t/ must satisfy a constraint on the statistical power expressed as Z C1 Ps . f / d f V P (13.1)
0
where Ps . f / is the power spectral density of s.t/. The signal-to-noise ratio of the channel as a function of frequency is dened as 0Ch . f / D jGCh . f /j2 Pw . f / (13.2)
The transmission passband B, dened in (13.8), is intuitively given by the interval of frequencies characterized by large values of 0Ch . f /. First, we want to determine an expression of the capacity C[b=s] that extends (6.280), valid for an ideal AWGN channel, to the case of a dispersive channel. To this end we divide the passband B, having measure B, into N sub-bands Bi , i D 1; : : : ; N , of width 1 f D B=N , where 1 f is chosen sufciently small so that Ps . f / and 0Ch . f / are, to a rst approximation, equal to constants within the generic sub-band Bi , that is we assume 9 f i 2 Bi such that Ps . f / D Ps . f i / and 0Ch . f / D 0Ch . f i /, 8 f 2 Bi .
1000
The subchannel i has a capacity C[b=s];i given by (6.280), C[b=s];i 1 f Ps . f i / jGCh . f i /j2 D 1 f log2 1 C 1 f Pw . f i / (13.3)
The total capacity C[b=s] is obtained by summing the terms C[b=s];i , that is
N X i D1
C[b=s];i D 1 f
N X i D1
(13.4)
and letting 1 f tend to zero, to obtain C[b=s] Z jGCh . f /j2 D log2 1 C Ps . f / log2 [1 C Ps . f / 0Ch . f /] d f df D Pw . f / B B Z (13.5)
We now nd the optimum power spectral density Ps . f / that maximizes the capacity (13.5), under the constraint (13.1) and the condition Ps . f / 0. Applying the method of Lagrange multipliers, the optimum Ps . f / maximizes the integral Z flog2 [1 C Ps . f / 0Ch . f /] C Ps . f /g d f (13.6)
B
where is a Lagrange multiplier. Using the calculus of variations (see Appendix 8.A) we nd that the optimum PSD must satisfy the following condition: 0Ch . f / C ln 2 D 0 1 C Ps . f / 0Ch . f / Therefore the solution is given for f 0 by 8 1 < K Ps;opt . f / D 0Ch . f / :0 (13.7)
f 2B otherwise
(13.8)
where B D f f 2 [0; C1/: Ps;opt . f / > 0g is the passband that allows achieving the capacity in (13.5), and K is a constant such that (13.1) is satised with the equal sign. This result is due to Shannon, and is valid for non-ideal linear channels in the presence of additive Gaussian noise. The function Ps;opt . f / is illustrated in Figure 13.1 for a typical behavior of the function 0Ch . f /; as the channel impulse response is assumed real valued, for f < 0 we get Ps;opt . f / D Ps;opt . f /. In fact, it results that Ps . f / should assume large values (small values) at frequencies for which 0Ch . f / assumes large values (small values). From Figure 13.1 we note that if 1= 0Ch . f / is the prole of a cup in which we pour a quantity of water equivalent to V P , the distribution of the water in the cup takes place according to the behavior predicted by (13.8); this observation leads to the water pouring interpretation of the optimum distribution of Ps . f / as a function of frequency.
1001
VP = Ps,opt (f) df
B
Ps,opt (f) =K K
1 (f) Ch 1 (f) Ch
Figure 13.1. Illustration of Ps,opt .f/ for a typical behavior of the function 0Ch .f/.
Shannon also demonstrated that capacity is achieved if s.t/ is a Gaussian process with power spectral density Ps;opt . f /; the capacity in bits per second is given by Z C[b=s] D
B
(13.9)
where B is dened by (13.8). By analogy with the case of an ideal (non-dispersive) AWGN channel with limited bandwidth (see (6.280)), a linear dispersive channel can be roughly characterized by two R parameters: the bandwidth B D B d f and the effective signal-to-noise ratio 0eff , implicitly dened so that C[b=s] D B log2 .1 C 0eff / The comparison with (13.9) yields Z 1 ln[1 C Ps;opt . f /0Ch . f /] d f 0eff D exp B B (13.10)
(13.11)
Note that 1 C 0eff corresponds to the geometric mean of the function 1 C Ps;opt . f /0Ch . f / in the band B. By analogy with the case of an ideal AWGN channel with limited bandwidth analyzed in Chapter 6, it is useful to dene a normalized signal-to-noise ratio for a transmission system over a linear dispersive channel that operates at a rate of the encoder-modulator equal to R I , in bits per dimension, as N 0eff D 0eff 2R I 2 1 (13.12)
N where 0eff > 1 measures the gap that separates the system being considered from capacity. The passband B that allows achieving capacity represents the most important parameter of the spectrum obtained by water pouring. For many channels utilized in practice, the passband B is composed of only one frequency interval [ f 1 ; f 2 ], as illustrated in Figure 13.1; in other cases, for example, in the presence of high-power narrowband interference signals,
1002
B may be formed by the union of disjoint frequency intervals. In practice, we nd that the dependence of the capacity on Ps . f / is not as critical as the dependence on B; a constant power spectral density in the band B usually allows a system to closely approach capacity [1, 2]. Therefore the application of the water pouring criterion may be limited to the determination of the passband.
13.2
A method to approach capacity is directly suggested by water pouring: the passband B that allows achieving capacity can be subdivided into disjoint sub-bands with sufciently small bandwidth 1 f so that both 0Ch . f / and Ps;opt . f / are approximately constant in each of the resulting subchannels; then each subchannel can be modeled as an ideal AWGN channel with bandwidth 1 f . This procedure is characteristic of OFDM systems, as discussed in Chapter 9. The power to be assigned to a signal transmitted over a subchannel having center frequency f is given by Ps;opt . f /1 f . Then, with good approximation, the signal-to-noise ratio in the subchannel is given by Ps;opt . f /0Ch . f /, and the corresponding capacity is given by Observing that B Ps;opt . f / d f D V P , as 1 f tends to zero, the total power in all sub-bands approximates V P and the total capacity approximates C[b=s] , as indicated by (13.9). Indeed, to achieve capacity, powerful coding techniques must be applied so that transmission over each subchannel may take place at a bit rate close to the subchannel capacity C[b=s] . f /. R C[b=s] . f / D 1 f log2 .1 C Ps;opt . f / 0Ch . f // (13.13)
We recall the denitions of the signal-to-noise ratio gap to capacity 0 gap;dB for an uncoded MQAM system, with M 1 and Pe D 10 7 , and of coding gain G code [i] achieved by coding at Pe D 10 7 , given by (12.5) in dB. By including in (13.13) the parameters 0 gap;dB and G code [i], we nd that the number of bits per modulation interval that can be transmitted over subchannel i with Pe D 10 7 is given by b[i] D log2 .1 C 0[i] 10.G code [i ]
0 gap;dB /=10
(13.15)
The achievable bit rate is obtained by summing the values given by (13.15) over all N subchannels, and multiplying the result by the modulation rate 1=T D 1 f , that is Rb D
N 1 X b[i] bit/s T i D1
(13.16)
1003
In practice we resort to a technique called bit loading to determine the number of bits to be transmitted over each subchannel per modulation interval, under the constraints: 1) b[i] can take only a nite number of values determined by the signal constellations; and 2) the total transmitted power is xed [3, 4, 5, 6]. Note that the modulation interval increases as the number of sub-bands increases. To reduce the delay in the recovery of the information, coding is usually applied across the subchannels, see, for example, [7].
where fak g is modeled as a sequence of complex-valued i.i.d. symbols with Gaussian 2 distribution and variance a , and h T x .t/ denotes the transmit lter impulse response, with Fourier transform HT x . f /. The transmit lter is chosen such that the PSD of s.t/, given by (7.27) and (7.28), is equal to Ps;opt . f /, that is
2 a jHT x . f /j2 D Ps;opt . f C f 0 / 1. f C f 0 / 4T
(13.18)
At the receiver, the signal r.t/ is rst demodulated and ltered by a lter with impulse response g w that suppresses the signal components around 2 f 0 and whitens the noise.
1004
1 Figure 13.2. Equivalence between a continuous time system, (a) passband model, (b) baseband equivalent model with gw .t/ D p gw .t/, 2 and (c) a discrete-time system, for transmission over a linear dispersive channel.
1005
As an alternative we could use a passband whitening lter phase splitter that suppresses the signal components with negative frequency and whitens the noise in the passband, see Section 8.14.1. 1 Consider the baseband equivalent model of Figure 13.2b, where GC . f / D p GCh . f C 2 f 0 / 1. f C f 0 / and PwC . f / D 2Pw . f C f 0 / 1. f C f 0 /; we dene B0 D [ f 1 f0 ; f2 f0 ] (13.19)
as the new passband of the desired signal at the receiver. The whitening lter gw .t/ D 1 p g w .t/ has frequency response given by
2
jGw . f /j2 D
8 <
1 PwC . f / : 0
f 2 B0 elsewhere
(13.20)
From the scheme of Figure 8.20, the whitening lter is then followed by a matched lter g M with frequency response G M . f / D [HT x . f / GC . f / Gw . f /] (13.21)
M The cascade of whitening lter gw and MF g M yields the composite lter g RcF . f / with frequency response G MF . f / D H x . f / GC . f / jGw . f /j2 Rc T
H x . f / GC . f / T PwC . f /
(13.22)
The overall QAM pulse q at the output of the MF has frequency response given by Q. f / D HT x . f / GC . f / G MF . f / D Rc jHT x . f / GC . f /j2 PwC . f / (13.23)
Note that Q. f / has the properties of a PSD with passband B0 ; in particular, Q. f / is equal to the noise PSD Pw R . f / at the MF output, as Pw R . f / D PwC . f / jG MF . f /j2 D Rc jHT x . f / GC . f /j2 D Q. f / PwC . f / (13.24)
ai h k
C wk Q
(13.25)
where the coefcients h i D q.i T / are given by the samples of the overall impulse response q.t/, and (13.26) wk D w R .kT / D [wC .t 0 / g MF .t 0 /].t/ Q Rc
tDkT
1006
In general, the Fourier transform of the discrete-time response fh i g is given by C1 1 X ` (13.27) H. f / D Q f T `D 1 T In this case, because Q. f / is limited to the passband B0 with bandwidth B D 1=T , there is no aliasing; the function H. f /, periodic of period 1=T , is therefore equal to .1=T /Q. f / Q in the band B0 . As Pw R . f / D Q. f /, fwk g is a sequence of Gaussian noise samples with autocorrelation sequence frwk .n/g D h n . Note, moreover, that fh i g satises the Hermitian Q property, as H is real valued. We have thus obtained a discrete-time equivalent channel that can be described using the D transform as x.D/ D a.D/h.D/ C w.D/ Q (13.28)
where h.D/ has Hermitian symmetry. We now proceed to develop an alternative model of discrete-time equivalent channel with causal, monic, and minimum-phase response fQ.D/, and additive white Gaussian noise w.D/. In this regard, we recall the theorem of spectral factorization for discrete time systems (see page 53). If H. f / satises the PaleyWiener condition, then the function h.D/ can be factorized as follows: 2 Q 1 f 0 fQ.D/ (13.29) h.D/ D f D
2 Q Q H. f / D F . f / f 0 F . f /
(13.30)
where the function fQ.D/ D 1 C fQ1 D C is associated with a causal ( fQi D 0 for i < 0), Q monic and minimum-phase sequence fQi , and F . f / D fQ.e j2 fT / is the Fourier transform 2 is the geometric mean of H. f / over an interval of of the sequence f fQi g. The factor f 0 measure 1=T , that is Z 2 log f 0 D T log H. f / d f (13.31)
1=T
where logarithms may have any common base. Then (13.28) can be written as 1 1 2 C w0 .D/ f 0 fQ x.D/ D a.D/ f 0 fQ.D/ fQ D D
(13.32)
where w0 .D/ is a sequence of i.i.d. Gaussian noise samples with unit variance. Filtering 2 x.D/ by a lter having transfer function 1=[ f 0 fQ .1=D /], we obtain the discrete-time equivalent canonical model of the dispersive channel z.D/ D a.D/ fQ.D/ C w.D/ (13.33)
2 where w.D/ is a sequence of i.i.d. Gaussian noise samples with variance 1= f 0 . Equation (13.33) is obtained under the assumption that fQ.D/ has a stable reciprocal function, and
1007
hence fQ .1=D / has an anticausal stable reciprocal function; this condition is veried if h.D/ has no spectral zeros. However, to obtain the reciprocal of fQ.D/ does not represent a problem, as z.D/ can be indirectly obtained from the sequence of samples at the output of the WMF. The transfer function of the composite lter that consists of the whitening lter gw and the WMF has a transfer function given by
WMF GRc . f / D Q G MF . f / H . f / GC . f / F . f / Rc D Tx 2F . f / Q PwC . f / H. f / f0
(13.34)
The only condition for the stability of the lter (13.34) is given by the PaleyWiener criterion.
B0
(13.36)
Z D
2 log2 1 C a Q. f / d f T B0
Recall from (13.27) that H. f /, periodic of period 1=T , is equal to .1=T /Q. f / in the band B0 ; therefore using (13.31) for B D 1=T , the capacity C [b=s] and its approximation
1008
for large values of the signal-to-noise ratio 0 can be expressed as Z Z 2 2 2 2 C[b=s] D log2 .1 C a H. f // d f ' log2 .a H. f // d f D B log2 .a f 0 / (13.37)
1=T 1=T
Assume that the tail of the impulse response that causes ISI can be in some way eliminated, so that at the receiver we observe the sequence a.D/ C w.D/ rather than the sequence (13.33). The signal-to-noise ratio of the resultant ideal AWGN channel becomes 2 2 0ISI free D a f 0 ; thus, from (6.280) the capacity of the ISI-free channel and its approximation for large values of 0 become CISI
free [b=s] 2 2 2 2 D B log2 .1 C a f 0 / ' B log2 .a f 0 /
(13.38)
(13.39)
Price was the rst to observe that for large values of 0 we obtain (13.39), that is for high signal-to-noise ratios the capacity C[b=s] of the linear dispersive channel is approximately equal to the capacity of the ideal ISI-free channel obtained assuming that ISI can be in some way eliminated from the sequence x.D/; in other words, ISI does not signicantly contribute to the capacity [1].
13.3
The convolution u.D/ D a.D/ fQ.D/ that determines the output sequence in the absence Q of noise can be viewed as the transform of a vector by the channel matrix, that is u D Fa, where a and u are vectors whose components are given by the transmitted symbols and channel output samples, respectively. As the canonical response fQ.D/ is causal and monic, Q the matrix F is triangular with all elements equal to 1 on the diagonal and therefore has Q determinant equal to 1; thus it follows that the matrix F identies a linear transformation that preserves the volume between the input and output spaces. In other words, the channel Q matrix F transforms a hypersphere into a hyperellipsoid having the same volume and containing the same number of constellation points. Coding methods for linear dispersive channels [8, 9, 10, 11, 12, 13, 14] that yield high values of coding gain can be obtained by requiring that the channel output vectors in the absence of noise u are points of a set 30 with good properties in terms of Euclidean distance, for example, a lattice identied by integers. From the model of Figure 13.2c, an intuitive explanation of the objectives of coding for linear dispersive channels is obtained by considering the signal sequences a, u, w, and z as vectors with a nite number of comQ ponents. If the matrix F is known at the transmitter, the input vector a can be predistorted Q so that the points of the vector a correspond to points of a signal set F 1 3; the volumes Q 1 3/ in the input signal space are equal. In any V .3/ in the output signal space, and V .F case, the output channel vectors are observed in the presence of additive white Gaussian noise vectors w. If a detector chooses the point of the lattice 30 with minimum distance from the output vector we obtain a coding gain, relative to 3, as in the case of an ideal AWGN channel (see Section 6.10).
1009
Recall that to achieve capacity it is necessary that the distribution of the transmitted signal approximates a Gaussian distribution. We mention that commonly used methods of shaping, which also minimize the transmitted power, require that the points of the input constellation are uniformly distributed within a hypersphere [11]. Coding and shaping thus occur in two Euclidean spaces related by a known linear transformation that preserves the volume. Coding and shaping can be separately optimized, by choosing a method for predistorting the signal set in conjunction with a coding scheme that leads to a large coding gain for an ideal AWGN channel in the signal space where coding takes place, and a method that leads to a large shaping gain in the signal space where shaping takes place. In the remaining part of this chapter we focus on precoding and coding methods to achieve large coding gains for transmission over channels with ISI, assuming the channel impulse response is known.
13.3.1
Precoding is a method of pre-equalization that allows also for transmission over linear dispersive channels the objectives of coding illustrated in Section 6.10 for transmission over ideal AWGN channels. We now extend the method discussed in Appendix 7.A for partial response systems. Let fQ.D/ D 1 C fQ1 D C fQ2 D 2 C be the response of a discretetime equivalent channel, with ISI and additive noise, and assume that the canonical response of the channel fQ.D/ is known at the transmitter. Let fQ.D/ D 1 C D fQ0 .D/, and furthermore assume that for every pair of symbols of the input constellation i ; j 2 A the following relation holds:1 i D j mod 30 , where A 30 C is a nite set of symbols (constellation) to which the information bits are mapped, 30 represents the lattice associated with A, and is a given offset value, possibly non-zero (see Section 12.1.3). The objective of all precoding methods, with and without channel coding, consists of transmitting a pre-equalized sequence a . p/ .D/ D u . p/ .D/ fQ.D/ (13.40)
so that, in the absence of noise, the channel output sequence u.D/ D a . p/ .D/ fQ.D/ D u . p/ .D/ represents the output of an ideal channel apparently ISI-free, with input sequence u . p/ .D/; u . p/ .D/ is a sequence of symbols belonging to a set A. p/ 30 C . To achieve this objective with channel input signal samples that belong to a given nite region, the cardinality of the set A. p/ must be greater than that of the set A. The redundancy in u . p/ .D/ can therefore be used to minimize the average power of the sequence a . p/ .D/ given by (13.40), or to obtain other desirable characteristics of a . p/ .D/, for example, a low peak-to-average ratio. In the case of systems that adopt trellis coding, the channel output sequence in the absence of noise u.D/ must be a valid code sequence and can then be decoded by a decoder designed for an ideal channel. Note that, in a system with precoding, the elements of the transmitted sequence a . p/ .D/ are not in general symbols with discrete values.
1
The expression i D j mod 30 denotes that the two symbols i and j differ by a quantity that belongs to 30 .
1010
The rst precoding method was independently proposed for uncoded systems by Tomlinson [15] and Harashima [16] (TH precoding). Initially, TH precoding was not used in practice because in an uncoded transmission system the preferred method to cancel ISI employs a DFE, as it does not require to send information on the channel impulse response to the transmitter. However, if trellis coding is adopted, decision-feedback equalization is no longer a very attractive solution, as reliable decisions are made available by the Viterbi decoder only with a certain delay. TH precoding, illustrated in Figure 13.3, uses memoryless operations at the transmitter and at the receiver to obtain samples of both the transmitted sequence a . p/ .D/ and the detected sequence a.D/ within a nite region that contains A. In principle, TH precoding O can be applied to arbitrary symbol sets A; however, unless it is possible to dene an efcient extension of the region containing A, the advantages of TH precoding are reduced by the increase of the transmit signal power (transmit power penalty). An efcient extension exists only if the signal space of a . p/ .D/ can be tiled, that is completely covered without overlapping with translated versions of a nite region containing A, given by the union of the Voronoi regions of symbols of A, and dened as R.A/. Figure 13.4 illustrates the efcient extension of a two-dimensional 16-QAM constellation, where 3T denotes the sublattice of 30 that identies the efcient extension. With reference to Figure 13.3, the precoder computes the sequence of channel input signal samples a . p/ .D/ as a . p/ .D/ D a.D/ where the sequence p.D/ D [ fQ.D/ 1] a . p/ .D/ D D fQ0 .D/ a . p/ .D/ (13.42) p.D/ C c.D/ (13.41)
represents the ISI at the channel output that must be compensated at the transmitter. The elements of the sequence c.D/ are points of the sublattice 3T used for the efcient extension of the region R.A/ that contains A.2 The k-th element ck 2 3T of the sequence c.D/ is . p/ chosen so that the statistical power of the channel input sample ak is minimum; in other . p/ words, the element ck is chosen so that ak belongs to the region R D R.A/, as illustrated in Figure 13.4. From (13.40), (13.41), and (13.42), the channel output sequence in the
bitmapper precoder discretetime channel (p) inverse bitmapper
b(D)
b a
a a(D) A
a (D) +
+ u(D) + + + w(D)
^ b(D)
Equation (13.41) represents the extension of (7.198) to the general case, in which the operation mod M is substituted by the addition of the sequence c.D/.
1011
point of lattice T =2L Z 2 L 0 0 Voronoi region ( 0 ) point of lattice 0 signal region R=R(A ) a A, L L signal set
Figure 13.4. Illustration of the efcient extension of a two-dimensional 16-QAM constellation p (L D M D 4).
absence of noise is given by u.D/ D a . p/ .D/ fQ.D/ D a . p/ [1 C D fQ0 .D/] D a . p/ .D/ C p.D/ D a.D/ C c.D/ (13.43)
Note that from (13.43) we get the relation u k D ak mod 3T , which is equivalent to (7.201). The samples of the sequence a . p/ .D/ can be considered, with a good approximation, uniformly distributed in the region R. Assuming a constellation with M D L L points for a QAM system, the power of the transmitted sequence is equal to that of a complex-valued signal with both real and imaginary parts that are uniformly distributed in [ .L=2/ 10 ; .L=2/ 10 ], where 10 denotes the minimum distance between points of the lattice 30 . Therefore using (5.34) it follows that L2 2 1 (13.44) 12 0 Recalling that the statistical power of a transmitted symbol in a QAM system is given by 2 2 L 12 1 12 (see (6.182) for 10 D 2), we nd that the transmit power penalty in a system 0 that applies TH precoding is equal to 12 =12 per dimension. 0 From (13.33), the channel output signal is given by E[jak j2 ] ' 2
. p/
(13.45)
where w.D/ represents a sequence of additive white Gaussian noise samples. In the case of TH precoding for an uncoded system, the detector yields a sequence u.D/ of symbols O belonging to the constellation A. p/ ; from (13.43) the detected sequence a.D/ of transmitted O symbols is therefore given by the memoryless operation a.D/ D u.D/ O O c.D/ O (13.46)
1012
The k-th element ck 2 3T of the sequence c.D/ is chosen so that the symbol ak D u k ck O O O O O belongs to the constellation A. As the inverse operation of precoding is memoryless, error propagation at the receiver is completely avoided. Moreover, as the inversion of fQ.D/ is not required at the receiver, fQ.D/ may exhibit spectral nulls, that is it can contain factors of the form .1 D/.
13.3.2
must be a valid code sequence. In practice, for all trellis codes that use a one-dimensional constellation with L points or a two-dimensional constellation with L L points, if L is a multiple of 4 and a.D/ is a code sequence, then also u.D/ is a valid code sequence; in this way trellis coding and TH precoding can be combined by applying to the encoded symbol sequence TH precoding as discussed in the previous section [11]. For example, Figure 13.5 illustrates
A =L Lpoint signal set (L=4) A 0 + offset ( 2 ) 0 y (0) B0 B1 1 + offset ( 2 ) 1 y (1)
1013
the partitioning of a 16-QAM constellation for trellis coding combined with TH precoding; note that summing a point of a lattice 3T (see Figure 13.4) with one of the points of the subset obtained at the partitioning level m C 1 D 2, we still obtain a point of the subset. Q In the general case of a one-dimensional constellation A with L points or a twodimensional constellation with L L points, with L even, the combination of trellis coding with TH precoding requires the application of trellis coding with feedback, or trellis augmented precoding, that we will now discuss [14]. Note that for L-PAM and L L-QAM constellations, the existence of an efcient extension is immediately veried. Moreover, for L even, we have that the subsets B0 31 C 0 and B1 31 C 1 , obtained at the rst 1 1 level of partitioning, are congruent through a translation dened by a point of 3T , with the sets of points that are again subsets of 31 C 0 and 31 C 1 , respectively. However, 1 1 this property is not necessarily veried for all partitioning levels up to level m C 1; an Q example is given by subsets obtained at the partitioning level m C 1 D 3 of a 6 6QAM Q constellation (see Figure 13.9). A system using feedback to combine trellis coding with TH precoding is illustrated in Figure 13.6. The state of the trellis code sk 1 is known at the transmitter. The symbol ak , into which the information represented by the binary vector bk is mapped, is taken from the set B y .0/ , that is one of the two subsets obtained at the rst level of partitioning.
.0/ The set B y .0/ is specied by the value yk D 0 or 1 of the element with index k of the
k
sequence y .0/ .D/, which is composed of the least signicant bits of the vector sequence that . p/ describes the evolution of the state of the trellis code. The output sample ak is determined by (13.41). As previously mentioned, to allow correct decoding operations the sequence u.D/ must represent a valid code sequence or, in other words, at the instant k the symbol u k must represent a valid continuation of the code sequence u.D/ starting from the code state sk 1 . The code sequence u.D/ is reproduced at the transmitter and presented at the input of a unit that determines, using the knowledge of the state sk 1 and the symbol u k , the next state sk ; this unit determines the bit sequence y .0/ .D/, such that the elements of the sequence a.D/ are chosen so that in turn u.D/ is a valid code sequence. The code sequence u.D/, received in the presence of additive white Gaussian noise (see (13.45)), is input to a Viterbi decoder that yields the detected symbol sequence u.D/. A O
precoder
b(D)
b a
a a(D) B
y (0)
+ + +
a (D) +
+ u(D) + + + w(D)
Viterbi decoder
^ u(D) +
^ a(D)
inverse bitmapper
^ b(D)
^ c(D)
y (0)
feedback TCM encoding
u(D)
1014
detection a.D/ of the sequence a.D/ is obtained by the memoryless operation (13.46), thus O O avoiding error propagation. A detection b.D/ of the binary information vector sequence b.D/ is then obtained by the inverse bit-mapping operation performed at the transmitter. An example of application of trellis augmented precoding will be given next. Example 13.3.1 Consider the transmission system with trellis augmented precoding illustrated in Figure 13.7. The code is an 8 state trellis code and the symbol constellation A is a 6 6-QAM constellation. Assume that the channel frequency response exhibits spectral nulls at f D 0 and f D 1=.2T / Hz, and is given by 1 fQ.D/ D 1 D2 D (13.48)
where 0 1. Figure 13.8a shows a conventional encoder for an 8-state trellis code that uses a systematic convolutional encoder for a code with rate 2/3 followed by a bit-mapper, and Figure 13.8b illustrates the code trellis diagram (see Chapter 12). The two-dimensional constellation A with 6 6 points and the set partitioning that yields the signal subsets assigned to the transitions on the trellis diagram are illustrated in Figure 13.9. .5/ .1/ The mapping of the information bits b D .bk ; : : : ; bk / 2 f.00000/; .00001/; : : : , .0/ .10001/g, alphabet of cardinality 18, to symbols ak 2 B y .0/ , where yk 2 f0; 1g, is illustrated k in Figure 13.9, where the lattice 31 , which will be used in the following, is also shown. In particular, we show in Figure 13.10 the representation of the binary vector yk D .2/ .1/ .0/ .yk ; yk ; yk /, obtained using the set partitioning of Figure 13.9. Furthermore, we choose for the symbols u k the representation given by Uk D Uk;I ; Uk;Q 2 2 Z2 C .1; 1/ (13.49)
bitmapper
precoder
discretetime channel
b(D) y (0)(D) D
b (a
a a(D) B (0))
y
a (D) +
(p)
u(D) + ~ f (D)
+ +
Viterbi decoder
^ u(D) +
^ a(D)
inverse bitmapper
^ b(D)
p(D) y (1)(D)
c(D)
w(D) D D
2 ~ f(D)= 1D 1D
^ c(D)
+ y (2) (D)
+ D u(D)
Mu > y
1015
y (1)
k (0) (0) sk1 =yk
(b)
Figure 13.8. Illustration of (a) conventional encoder for an 8-state trellis code and (b) trellis diagram.
that is u k;I D u k;Q D 8u .3/ C 4u .2/ C 2u .1/ C 1 C ck k;I k;I k;I 8u .3/ C 4u .2/ C 2u .1/ C 1 C ck k;Q k;Q k;Q (13.50)
1016
u2 A
y y
(1) (0)
+b
b b b
b y
(1) (0)
y y 4/01 4/00 4/10 5/01 5 5/10 6/01 6/10 4/11 5/00 3 5/11 6/00 6/11
(1) (0)
5 3 1 1 3 5
u1
B0
y (0) + 0
y (0) + 1
B1
y (1) + 0 +4
y (1) + 1
Lattice L 1 C1 C3
C0
3 0 7 2 1 8
C2
5 6
y (2) + 0
y (2) + 1
D0
D4
D2
D6
D1
D5
D3
D7
with u .i / ; u .i / 2 f0; 1g, and ck 2 3T . Note that the symbols ak have a binary representation k;I k;Q obtained by setting ck D 0 in (13.50). It is possible to express the components of the vector yk as a function of u .i / ; u .i / , i D 1; 2, using for example the Karnaugh maps, as k;Q k;Q
.0/ yk D u .1/ u .1/ k;I k;Q .1/ yk D u .1/ k;Q .2/ yk
u .2/ k;I
1017
u Q , aQ
y (2) y (1) y (0) 001 000 101 5 010 011 110 3 101 5 110 100 3 111 001 1 1 010 1 101 3 110 5 1 011 3 110 5 111 000 101 100 111 010 011 100 001 000
uI , a I
001
000
100
001
000
010
011
111
010
011
The sequence of channel input samples is given by (13.41), where p.D/ D [ fQ.D/ 1] a . p/ .D/ D D D . p/ a .D/ 1 D (13.52)
The elements of the sequence c.D/ are points of the lattice 3T that determines the efcient extension of R.A/, as illustrated in Figure 13.11. Recall that the value of the el. p/ ement ck 2 3T is chosen so that the statistical power of the channel input sample ak is minimum. The symbol sequence u.D/, which represents a valid trellis code sequence at the output of the channel with response fQ.D/ in the absence of noise, is given by (13.47). At each instant k the element u k of the sequence u.D/ is presented at the input of the unit that determines the next state of the code sequence starting from the present state .1/ .2/ sk 1 . The state sk is obtained by rst determining the binary elements yk and yk of the .1/ .D/ and y .2/ .D/, as indicated by (13.51) and (13.50). Then the elements sequences y .1/ .2/ yk and yk are input to the systematic convolutional encoder for the code with rate 2/3; .0/ the encoder determines the next state sk and it outputs the bit ykC1 , so that the map can generate a symbol akC1 that is a valid continuation of the code sequence. At the receiver, the Viterbi decoder outputs the sequence u.D/. A detection a.D/ of the O O sequence a.D/ is given by the memoryless operation a.D/ D u.D/ O O c.D/ O (13.53)
Then the detected sequence of information bits is obtained from the sequence a.D/. O
1018
0 =2
a B0
a B1
Note that the assumption of perfectly known channel characteristics only holds in an ideal situation. For example, if the considered method is applied to high speed data transmission over UTP cables, we recall from Section 4.4.2 that low-frequency disturbances and near-end cross-talk at high frequency are the main impairments. In this case it is not practical to convey to the transmitter information about the channel. The overall system must therefore be designed for worst-case channel characteristics, and deviations from the assumed characteristics must be compensated at the receiver by adaptive equalization.
13.3.3
Flexible precoding
Flexible precoding was originally introduced during the development of the V.34 modem standard for data transmission over the PSTN up to 28.8 kbit/s; a further version was applied in the V.34 bis standard for data transmission up to 33.6 kbit/s [12, 13]. First version. In the rst version, illustrated in Figure 13.12, the transmitted signal a . p/ .D/ can be expressed as a . p/ .D/ D a.D/ C d.D/ where d.D/ is called the dither signal and is given by d.D/ D c.D/ p.D/ (13.55) (13.54)
where p.D/ is given by (13.42) and c.D/ is obtained from the quantization of p.D/ with quantizer Q 3mC1 . The quantizer Q 3mC1 yields the k-th element of the sequence c.D/ by Q Q
1019
precoder
b(D)
b a
a a(D) A
a (D)
+ u(D) + + + w(D)
Viterbi decoder
^ u(D)
~ Df (D)
^ b(D)
~ m+1
c(D)
^ c(D)
quantizing the sample pk to the closest point of the lattice 3mC1 , which corresponds to the Q (m C 1)-th level of partitioning of the signal set (see Chapter 12); in the case of an uncoded Q sequence, it is 3mC1 D 30 . Note that the dither signal can be interpreted as the signal Q with minimum amplitude that must be added to the sequence a.D/ to obtain a valid code sequence at the channel output in the absence of noise. In fact, at the channel output we get the sequence z.D/ D u.D/ C w.D/, where u.D/ is obtained by adding a sequence of points taken from the lattice 3mC1 to the code sequence a.D/, and therefore it represents Q a valid code sequence. The sequence z.D/ is input to a Viterbi decoder, which yields the detected sequence u.D/. To obtain a detection of the sequence a.D/, it is rst necessary to detect the sequence O c.D/ of the lattice points 3mC1 added to the sequence a.D/. Observing O Q a . p/ .D/ D O and O p.D/ D D fQ0 .D/ a . p/ .D/ O (13.57) u.D/ O fQ.D/ (13.56)
the sequence c.D/ is obtained by quantizing with a quantizer Q 3mC1 the sequence p.D/, O O Q as illustrated in Figure 13.12. Then subtracting the sequence c.D/ from u.D/ we obtain O O the sequence a.D/, that is used to detect the sequence of information bits. O At this point we can make the following observations with respect to the rst version of exible precoding: 1. an efcient extension of the region R.A/ is not necessary; 2. it is indispensable that the implementation of the blocks that perform similar functions in the precoder and in the inverse precoder (see Figure 13.12) is identical with regard to the binary representation of input and output signals; 3. as the dither signal can be assumed uniformly distributed in the Voronoi region V .3mC1 / of the point of the lattice 3mC1 corresponding to the origin, the transmit Q Q power penalty is equal to 12Q =12 per dimension; this can signicantly reduce the mC1 coding gain if the cardinality of the constellation A is small;
1020
4. to perform the inverse of the precoding operation, the inversion of the channel transfer function is required (see (13.56)); if fQ.D/ is minimum phase then 1= fQ.D/ is stable and the effect of an error event at the Viterbi decoder output vanishes after a certain number of iterations; on the other hand, if fQ.D/ has zeros on the unit circle (spectral nulls), error events at the Viterbi decoder output can result in an unlimited propagation of errors in the detection of the sequence a.D/. Second version. The second version of exible precoding, illustrated in Figure 13.13, includes trellis coding with feedback, as previously discussed. In the precoder and in the inverse precoder the quantizer Q 31 is now used, that yields the element ck by quantizing the sample pk to the closest point of the lattice 31 . Note that, if the symbol ak 2 B y .0/ , where
.0/ we recall that yk represents the least signicant bit of the trellis code state vector sk 1 , then we have u k D ak Cck 2 B y .0/ . In the coding method with feedback from the knowledge k of the state sk 1 and the symbol u k , the next state sk is determined, and consequently also .0/ the bit ykC1 ; at the channel output, in the absence of noise, we therefore obtain a valid code sequence u.D/. Note that, as the dither signal is now uniformly distributed in the region V .31 /, the transmit power penalty is reduced to 12 =12 per dimension (see Figure 13.9). 1 The problem of unlimited error propagation that occurs in exible precoding if the frequency response of the channel exhibits spectral nulls can be mitigated by referring to the scheme illustrated in Figure 13.14. Consider the signals and regions for a two-dimensional constellation A as illustrated, for example, in Figure 13.15 for a 16-QAM constellation; assume that the transfer function of the channel, that can exhibit spectral nulls, has the form
k
(13.58)
To set a limit to error propagation, we exploit the knowledge that the expression of the . p/ dither signal is such that the transmitted signal sample at instant k, ak , must be conned within the region R y .0/ , obtained by the union of Voronoi regions of points of the subset
k
precoder
b(D)
b a
a a(D) B
y (0)
a (D)
+ u(D) + + + w(D)
Viterbi decoder
^ u(D)
~ Df (D)
^ b(D)
Q ^ c(D)
y (0)
feedback TCM encoding
c(D)
Figure 13.13. Block diagram of a system with exible precoding and trellis coding with feedback.
1021
precoder
b(D)
b a
a a(D) B
y (0)
+ u(D) +
inverse precoder
a (D)
F
p(D) Q
1
+ u(D) + + + w(D)
Viterbi decoder
F
^ p(D) Q
1
^ a(D)
inverse bitmapper
^ b(D)
y (0) (D)
~ Df (D) c(D)
y (0)
feedback TCM encoding
^ c(D)
F
^ u(D) ^ (0) (D) y
~ Df 0 (D)
lim. to
~ f(D)
Ry (0)
+ +
~ DfN (D)
Figure 13.14. Block diagram of a system with exible precoding and trellis coding with feedback that includes a method to mitigate error propagation in the inverse precoder.
a B0 A= B 0 a B1 B1
signal region R 0
signal region R 1
Figure 13.15. Illustration of signals and signal regions of a 16-QAM constellation for a system with exible precoding and mitigation of the error propagation in the inverse precoder.
1022
B y .0/ , dened as
k
R y .0/ D fak
k
. p/
D C d : 2 B y .0/ ; d 2 V .31 /g
k
.0/ yk D 0; 1
(13.59)
In the precoder and in the inverse precoder, we use identical units denoted as F (see Figures 13.14 and 13.15). Considering, for example, the transmitter we see that unit F . p/ contains a non-linear element that limits the transmitted sequence sample ak to the region .0/ .0/ R y .0/ , selected as R0 if yk D 0, or as R1 if yk D 1. Note that the output signal of the k non-linear element is input of the recursive section of a lter: this section is used for the inversion of the numerator of fQ.D/, which determines the presence of spectral nulls in the transfer function of the channel. Example 13.3.2 Consider a dispersive AWGN channel with transfer function given by (13.48) with D 7=8, and a QAM transmission system with a 6 6 constellation for an 8-state trellis code. The frequency response of the channel and the constellation A are illustrated in Figures 13.16 and 13.9, respectively. Recall that the frequency response of the channel exhibits spectral nulls at f D 0 and f D 1=.2T / Hz. Figure 13.17 illustrates the curves of symbol error probability as a function of the signalto-noise ratio 0 obtained by simulating systems that employ: a) TH precoding and trellis coding with feedback; b) exible precoding and trellis coding with feedback and limitation of error propagation; c) uncoded 18-QAM and a receiver with DFE. Note that, for the considered AWGN channel and a transmission of log2 .18/ bit/s/Hz with an 8-state trellis code and 6 6-QAM, the system a) offers a margin larger than 1 dB with respect to system b), and a margin of 3 dB with respect to an uncoded system. This result can be explained by recalling that the error propagation is completely eliminated with TH precoding, while it can only be mitigated with exible precoding.
10 Amplitude (dB) 0 10 20 30 40
1/2 fT
Figure 13.16. Frequency response with spectral nulls for f D 0 and f D 1=.2T/ Hz.
1023
Figure 13.17. Symbol error probability as a function of the signal-to-noise ratio for systems that use: (a) TH precoding combined with trellis coding with feedback, (b) exible precoding combined with trellis coding with feedback and limitation of error propagation, (c) uncoded 18-QAM and receiver with DFE.
trellis coding & bitmapper precoder discretetime channel (p) inverse precoder inverse
b(D)
b a
a a(D) B
z (0)
a (D)
+ u(D) + + + w(D)
Viterbi decoder
^ u(D)
z (0)
~ Df (D)
Q ^ c(D)
y (0)
feedback TCM encoding
c(D)
Figure 13.18. Block diagram of a system with exible precoding and trellis coding with feedback, and a transmit power penalty equal to 12 =12 per dimension. 0
A further reduction of the transmit power penalty in the exible precoding and trellis coding with feedback can be obtained by using the scheme shown in Figure 13.18, where the quantizer Q 30 yields the element ck by quantizing the sample pk to the closest point of the lattice 30 . Note that the dither signal is now uniformly distributed in V .30 / and therefore the transmit power penalty is reduced to 12 =12 per dimension, that is the same value 0 obtained with TH precoding. Observe that we have ck 2 31 or ck 2 31 C .1; 0/. To obtain a valid code sequence u.D/, we recall that it is necessary to obtain u k D ak C ck 2 B y .0/ .
k
1024
precoder
discretetime channel
inverse precoder
b(D)
b a
a a(D) B
z (0)
+ u(D) +
a (D)
(p)
F
p(D) Q
0
+ u(D) + + + w(D)
^ u(D)
Viterbi decoder
F
^ (0) (D) z Q
0
inverse bitmapper
^ b(D)
z (0)
z (0) (D)
^ y (0) (D)
~ Df (D) c(D)
y (0)
feedback TCM encoding
F
^ u(D) ^ (0) (D) y
~ Df 0 (D)
lim. to
~ f(D)
Ry (0)
+ +
~ DfN (D)
Figure 13.19. Method of mitigating error propagation for the system of Figure 13.18.
a B0 A= B 0 a B1 B1
signal region R 0
signal region R 1
Figure 13.20. Illustration of signals and signal regions of a 16-QAM constellation for the system of Figure 13.19.
13. Bibliography
1025
z k D yk if ck 2 31 C .1; 0/, where yk denotes the 1s complement of yk . N N However, note that also with this version of exible precoding channel inversion is required for the precoding inverse operation (see (13.56)). In case the channel frequency response has spectral nulls, mitigation of error propagation in detecting the sequence a.D/ is obtained by the scheme illustrated in Figure 13.19. Regions used to limit in the units F O the sequence of transmitted samples a . p/ .D/ and the sequence of detected samples a . p/ .D/ are shown in Figure 13.20 for a 16-QAM constellation.
.0/
.0/
.0/
.0/
Bibliography
[1] R. Price, Nonlinearly feedback-equalized PAM versus capacity for noisy lter channels, Proc. 1972 Int. Conf. Comm., pp. 22.1217, June 1972. [2] I. Kalet, The multitone channel, IEEE Trans. on Communications, vol. 37, pp. 119 124, Feb. 1989. [3] J. A. C. Bingham, Multicarrier modulation for data transmission: an idea whose time has come, IEEE Communications Magazine, vol. 28, pp. 514, May 1990. [4] P. S. Chow, J. M. Ciof, and A. C. Bingham, A practical discrete multitone transceiver loading algorithm for data transmission over spectrally shaped channels, IEEE Trans. on Communications, vol. 43, pp. 773775, Feb./March/April 1995. [5] A. Leke and J. M. Ciof, A maximum rate loading algorithm for discrete multitone systems, in Proc. Globecom 1997, pp. 15141518, Nov. 1997. [6] J. Campello, Practical bit loading for DMT, in Proc. IEEE International Conference on Communications, Vancouver, Canada, pp. 801805, June 1999. R [7] G. Cherubini, E. Eleftheriou, and S. Olcer, Filtered multitone modulation for veryhigh-speed digital subscriber lines, IEEE Journal on Selected Areas in Communications, June 2002. [8] J. M. Ciof, G. P. Dudevoir, M. V. Eyuboglu, and G. D. Forney, Jr., MMSE decisionfeedback equalizers and coding, IEEE Trans. on Communications, vol. 43, pp. 2582 2604, Oct. 1995. [9] G. Ungerboeck, Channel coding with multilevel/phase signals, IEEE Trans. on Information Theory, vol. 28, pp. 5567, Jan. 1982. [10] G. D. Forney, Jr. and G. Ungerboeck, Modulation and coding for linear Gaussian channels, IEEE Trans. on Information Theory, vol. 44, pp. 23842415, Oct. 1998.
1026
[11] M. V. Eyuboglu and G. D. Forney, Jr., Trellis precoding: combined coding, precoding and shaping for intersymbol interference channels, IEEE Trans. on Information Theory, vol. 38, pp. 301314, Mar. 1992. [12] R. Laroia, S. A. Tretter, and N. Farvardin, A simple and effective precoding scheme for noise whitening on intersymbol interference channels, IEEE Trans. on Communications, vol. 41, pp. 14601463, Oct. 1993. [13] R. Laroia, Coding for intersymbol interference channelsCombined coding and precoding, IEEE Trans. on Information Theory, vol. 42, pp. 10531061, July 1996. [14] G. Cherubini, S. Olcer, and G. Ungerboeck, Trellis precoding for channels with spectral nulls, in Proc. 1997 IEEE Int. Symposium on Information Theory, Ulm, Germany, p. 464, June 1997. [15] M. Tomlinson, New automatic equalizer employing modulo arithmetic, Electronics Letters, vol. 7, pp. 138139, Mar. 1971. [16] H. Harashima and H. Miyakawa, Matched transmission technique for channels with intersymbol interference, IEEE Trans. on Communications, vol. 20, pp. 774780, Aug. 1972.
Chapter 14
Synchronization
As a generalization of the receiver block diagram shown in Figure 7.5, the representation of the analog front end for a passband QAM system is illustrated in Figure 14.1. The received signal r.t/ is multiplied by a complex-valued carrier generated by a local oscillator, then ltered by an anti-imaging lter, g AI , that extracts the image of the signal yielding the I and Q components of the signal rC .t/. Often the function of the anti-imaging lter is performed by other lters, however, here g AI is considered only as a model for the analysis and is assumed to be non-distorting. The receive oscillator is independent of the transmit oscillator; consequently carrier recovery must be performed using one of the following two strategies. 1. The rst consists in multiplexing, usually in the frequency domain, a special signal, called a pilot signal, at the transmitter. This allows extracting the carrier at the receiver and therefore synchronizing the receive oscillator in phase and frequency with the transmit oscillator. If the pilot signal v.t/ consist of a non-modulated carrier, carrier recovery is obtained by the phase-locked loop (PLL) described in Section 14.2. 2. The second consists in getting the carrier directly from the modulated signal; this approach has the advantage that all the transmitted power is allocated for the transmission of the signal carrying the desired information. Some structures that implement this strategy are reported in Section 14.3. In this chapter we will discuss methods for carrier phase and frequency recovery, as well as algorithms to estimate the timing phase. To avoid ambiguity, we refer to the latter as timing recovery algorithms, dropping the term phase. These algorithms are developed for application in the PAM and QAM transmission systems of Chapter 7, and the spread spectrum systems of Chapter 10. The problem of synchronization of OFDM systems was mentioned in Section 9.7.
14.1
In Figure 14.1 the reconstructed carrier, in complex form, has the expression vC .t/ D expf j .2 f 1 t C '1 /g O (14.1)
1028
Therefore, if the carrier generated at the transmitter, as observed at the receiver (received carrier) is given by expf j .2 f 0 t C '0 /g,1 in general we have a reconstruction error of the carrier phase ' P A .t/ given by ' P A .t/ D .2 f 1 t C '1 / Let D 2. f 0 D .'0 then (14.2) can be rewritten as ' P A .t/ D t (14.4) '1 / f1/ (14.3) .2 f 0 t C '0 / (14.2)
With reference to the notation of Figure 7.12, observing that now the phase offset is not included in the baseband equivalent channel impulse response, we have 1 .bb/ gC .t/ D p gCh .t/ e 2 2
j arg GCh . f 0 /
(14.5)
The resulting baseband equivalent scheme of a QAM system is given in Figure 14.2. We assume that the anti-imaging lter frequency response is at within the frequency interval jfj B C where B is the bandwidth of the signal sC and
max
2
max
s (bb)(t)
g (t) C
sC (t)
(t)
PA
rC (t)
wC (t)
e-j
Figure 14.2. Baseband equivalent model of the channel and analog front end for a QAM system.
In this chapter '0 is given by the sum of the phase of the transmitted carrier and the channel phase at f D f 0 , equal to arg G Ch . f 0 /.
1029
The received signal rC is affected by a frequency offset and a phase offset ; moreover, the transmit lter h T x and the channel lter gC introduce a transmission delay t0 . To simplify the analysis, this delay is assumed to be known with an error that is in the range . T =2; T =2/. This coarse timing estimate can be obtained, for example, by a correlation method with known input (see (7.269)). This corresponds to assuming the overall pulse qC non-causal, with peak at the origin. Once set t0 D "T , with j"j 1=2, the signal rC .t/ can be written as rC .t/ D e j . where qC .t/ D .h T x gC /.t C "T / wC' .t/ D wC .t/ e
j' P A .t/ tC / C1 X kD 1
ak qC .t
kT
(14.7)
(14.8)
Furthermore, the receiver clock is independent of the transmitter clock, consequently the receiver clock period, that we denote as Tc , is different from the symbol period T at the transmitter: we assume that the ratio F0 D T =Tc is in general a real number. The synchronization process consists in recovering the carrier, in the presence of phase offset and frequency offset , and the timing phase or time shift "T .
14.2
We assume the transmitted sinusoidal component v.t/ has been isolated from the signal at the receiver by a suitable narrowband lter. Now the problem consists in generating by a local oscillator, for example, using a PLL, a signal vVCO .t/ with the same frequency and phase of v.t/, apart from a known offset. A PLL, whose block diagram is shown in Figure 14.3, is a control system used to automatically regulate the phase of a locally generated signal vVCO .t/ so that it coincides with that of an input signal v.t/. We assume that the two signals are given by v.t/ D A1 sin[!0 t C '0 .t/] vVCO .t/ D A2 cos[!1 t C '1 .t/] (14.9)
1030
where the phase '0 .t/ is a slowly time-varying function with respect to !0 , or d'0 .t/ dt !0 We write the instantaneous phase of vVCO .t/ as follows: !1 t C '1 .t/ D !0 t C .!1 and we let '0 .t/ D .!1 O thus vVCO .t/ D A2 cos[!0 t C '0 .t/] O !0 / t C '1 .t/ !0 / t C '1 .t/
(14.10)
(14.11)
(14.12)
(14.13)
and '0 .t/ then represents the estimate of the phase '0 .t/ obtained by the PLL. O We dene the phase error as the difference between the instantaneous phases of the signals v.t/ and vVCO .t/, that is .t/ D .!0 t C '0 .t// .!1 t C '1 .t// D '0 .t/ '0 .t/ O (14.14)
As illustrated in Figure 14.3, a PLL comprises: a phase detector (PD), that yields an output signal e.t/ given by the sine of the difference between the instantaneous phases of the two input signals, that is e.t/ D K D sin[ .t/] (14.15)
where K D denotes the phase detector gain; we observe that e.t/ is an odd function of .t/; therefore the PD produces a signal having the same sign as the phase error, at least for values of between and C ; a lowpass lter F.s/, called loop lter, whose output u.t/ is equal to u.t/ D f e.t/ (14.16)
a voltage-controlled oscillator (VCO), which provides a periodic output signal vVCO .t/ whose phase '0 .t/ satises the relation O d '0 .t/ O D K 0 u.t/ dt called VCO control law, where K 0 denotes the VCO gain. In practice, the PD is often implemented by a simple multiplier; then the signal e.t/ is proportional to the product of v.t/ and vVCO .t/. If K m denotes the multiplier gain and we dene KD D
1 2
(14.17)
A1 A2 K m
(14.18)
1031
then we obtain e.t/ D K m v.t/ vVCO .t/ D K m A1 sin[!0 t C '0 .t/] A2 cos[!1 t C '1 .t/] D K D sin[ .t/] C K D sin[2!0 t C '0 .t/ C '0 .t/] O Note that, with respect to the signal e.t/ dened in (14.15), there is now an additional term with radian frequency 2!0 . However, as from (14.10) '0 .t/ is slowly varying in comparison with the term at frequency 2!0 , the high frequency components are eliminated by the lowpass lter F.s/ or, in case the lowpass lter is not implemented, by the VCO that has a lowpass frequency response. Therefore the two schemes, with a PD or with a multiplier, may be viewed as equivalent; because of its simplicity, the latter will be considered in the following analysis. (14.19)
14.2.1
We now derive a baseband equivalent model of the PLL. From (14.16) we have Z t u.t/ D f .t / e. / d
0
(14.20)
f .t
/ e. / d
(14.21)
By this relation we derive the baseband scheme of Figure 14.4. Subtraction of the phase estimate '0 .t/ from the phase '0 .t/ yields the phase error .t/ that, transformed by the O non-linear block K D sin./, in turn gives the signal e.t/. The signal e.t/ is input to the loop lter F.s/, which outputs the control signal u.t/. The integration block, with gain K 0 , integrates the signal u.t/ and yields the estimate '0 .t/, thus closing the loop. O .t/ for '0 .t/, and of the expression O Substitution in (14.21) of the quantity '0 .t/ (14.15) for e.t/, yields Z t d .t/ d'0 .t/ D K D K0 f .t / sin[ . /] d (14.22) dt dt 0
1032
The (14.22) represents the integro-differential equation that governs the dynamics of the PLL. Later we will study this equation for particular expressions of the phase '0 .t/, and only for the case .t/ ' 0, i.e. assuming the PLL is in the steady state or in the so-called lock condition; the transient behavior, that is for the case .t/ 6D 0, is difcult to analyze and we refer to [1] for further study.
Linear approximation
Assume that the phase error tion holds .t/ is small, or .t/ ' 0; then the following approxima(14.23)
sin[ .t/] ' .t/ and (14.22) simplies into d'0 .t/ d .t/ D dt dt Z K D K0
0 t
f .t
/ . / d
(14.24)
In this way the non-linear block K D sin./ of Figure 14.4 becomes a multiplier by the constant K D , and the whole structure is linear, as illustrated in the simplied block diagram of Figure 14.5. We denote by P .s/ the Laplace transform of .t/; by taking the Laplace transform of (14.24) and assuming '0 .0/ D 0 we obtain O s P .s/ D s80 .s/ Substituting P .s/ with 80 .s/ H .s/ D K D K 0 F.s/ P .s/ (14.25)
O 80 .s/, we derive the loop transfer function as O 80 .s/ K F.s/ D 80 .s/ s C K F.s/ K D K D K0 (14.26)
Then from (14.26) we get the following two relations: P .s/D80 .s/ O 80 .s/ D [1 H .s/] 80 .s/ (14.27) (14.28)
1033
Table 14.1 Three expressions of '0 .t/ and corresponding Laplace transforms.
80 .s/ 's s !s s2 !r s3
We dene as steady state error 1 the limit for t ! 1 of .t/; recalling the nal value theorem, and using (14.28), 1 can be computed as follows:
1
D lim
t!1
1 1 C K F.s/=s
(14.29)
We compute now the value of 1 for the three expressions of '0 .t/ given in Table 14.1 along with the corresponding Laplace transforms. phase step: '0 .t/ D 's 1.t/;
1
(14.30)
thus we obtain
1
D 0 () F.0/ 6D 0
(14.31)
Observe that (14.31) holds even if F.s/ D 1, i.e. in case the loop lter is absent. frequency step: '0 .t/ D !s t 1.t/; !s !s 1 D lim 1 D lim s 2 s!0 s!0 s C K F.s/ s 1 C [K F.s/=s] If we choose F.s/ D s then
1 k
(14.32)
F1 .s/
(14.33)
D 0.
1034
frequency ramp: '0 .t/ D .!r t 2 =2/ 1.t/ !r 1 !r D lim 2 1 D lim s 3 s!0 s!0 s C K F.s/s s 1 C [K F.s/=s]
(14.34)
If we use a loop lter of the type (14.33) with k D 1, i.e. with one pole at the origin, then we obtain a steady state error 1 given by
1
!r 6D 0 K F1 .0/
(14.35)
As a general rule we can state that, in the presence of an input signal having Laplace transform of the type s k with k 1, to get a steady state error 1 D 0, a lter with at least .k 1/ poles at the origin is needed. The choice of the above elementary expressions of the phase '0 .t/ for the analysis is justied by the fact that an arbitrary phase '0 .t/ can always be approximated by a Taylor series expansion truncated to the second order, and therefore as a linear combination of the considered functions.
14.2.2
We now extend the PLL baseband model and relative analysis to the case in which white noise w.t/ with spectral density N0 =2 is added to the signal v.t/. Introducing the in-phase and quadrature components of w.t/, from (1.162) we get the relation w.t/ D w I .t/ cos.!0 t/ w Q .t/ sin.!0 t/ (14.36)
where w I .t/ and w Q .t/ are two uncorrelated random processes having spectral density in the desired signal band given by Pw I . f / D Pw Q . f / D N0 Letting Kw D
1 2
(14.37)
A2 K m
(14.38)
the multiplier output signal e.t/ assumes the expression e.t/ D K m [v.t/ C w.t/] vVCO .t/ D K D sin[ .t/] C K w w Q .t/ sin['0 .t/] C K w w I .t/ cos['0 .t/] O O C K D sin[2!0 t C '0 .t/ C '0 .t/] O C K w w I .t/ cos[2!0 t C '0 .t/] O Ignoring the high-frequency components in (14.39), (14.20) becomes Z t u.t/ D f .t /fK D sin[ . /] C K w w Q . / sin['0 . /] C K w w I . / cos['0 . /]g d O O
0
(14.39)
(14.40)
1035
Dening the noise signal we .t/ D K w [w I .t/ sin '0 .t/ C w Q .t/ cos '0 .t/] O O (14.41)
from (14.21) we get the integro-differential equation that describes the dynamics of the PLL in the presence of noise, expressed as Z t d .t/ d'0 .t/ f .t /fK D sin[ . /] C we . /g d (14.42) D K0 dt dt 0 From (14.42) we obtain the PLL baseband model illustrated in Figure 14.6.
Figure 14.7. Linearized PLL baseband model in the presence of additive noise.
1036
where H. f / D H . j2 f / (14.45)
To obtain Pwe . f / we use (14.41). Assuming w I .t/ and w Q .t/ are uncorrelated white random processes with autocorrelation rw I . / D rw Q . / D N0 . / (14.46)
and using the property of the Dirac function . / f .t C / D . / f .t/, the autocorrelation of we .t/ turns out to be rwe .t; t
2 / D K w rw I . /fsin['0 .t/] sin['0 .t O O
/]g (14.47)
(14.48)
Therefore using (14.18) and (14.38), from (14.44) we get the variance of the phase error, given by Z Z C1 1 N0 C1 (14.49) jH. f /j2 Pwe . f / d f D 2 jH. f /j2 d f 2 D 2 KD A1 1 1 From (1.139) we now dene the equivalent noise bandwidth of the loop lter as Z C1 jH. f /j2 d f 0 (14.50) BL D jH.0/j2 Then (14.49) can be written as 2 D 2N0 B L A2 1 (14.51)
where A2 =2 is the statistical power of the desired input signal, and N0 B L D .N0 =2/2B L is 1 the input noise power evaluated over a bandwidth B L . In Table 14.2 the expressions of B L for different choices of the loop lter F.s/ are given.
14.2.3
In this section we analyze the behavior of a second-order PLL, using the linearity assumption. In particular, we nd the expression of the phase error .t/ for the input signals given in Table 14.1, and we evaluate the variance of the phase error 2 .
1037
Table 14.2 Expressions of BL for different choices of the loop lter F.s/.
BL K 4 K Ca 4 K .K C a/ 4.K C b/ K .a K C a 2 b/ 4.a K b/
s 2 C as C b K .s 2 C as C b/ s2 s 3 C K s 2 C a K s C bK
From Table 14.2 the transfer function of a second-order loop is given by H .s/ D K .s C a/ s2 C K s C K a (14.52)
We dene the natural radian frequency, !n , and the damping factor of the loop, , as !n D p Ka D 1 p Ka 2a (14.53)
(14.54)
Once the expression of '0 .t/ is known, .t/ can be obtained by (14.27) and nding the inverse transform of P .s/. The relation is simplied if in place of s we introduce the normalized variable s D s=!n ; in this case we obtain Q s P .Q / D s2 Q 8 .Q / s 2C2 sC1 0 s Q Q (14.55)
which depends only on the parameter . In Figures 14.8, 14.9, and 14.10 we show the plots of .t/, with as a parameter, for the three inputs of Table 14.1, respectively. Note that for the rst two inputs .t/ converges to zero, while for the third input it converges to a non-zero value (see (14.35)), because F.s/ has only one pole at the origin, as can be seen from Table 14.2. We note that if .t/ is a phase step the speed of convergence increases with increasing , whereas if .t/ is a frequency step the speed of convergence is maximum for D 1. In Figure 14.11 the plot of B L is shown as a function of . As 1 B L D !n C (14.56) 2 8
1038
0.8
0.6
0.4
() s
0.2
=5 =2
0.2
=1 =0.7 =0.5
0.4
10
12
=n t
Figure 14.8. Plots of . / as a function of D !n t, for a second-order loop lter with a phase step input signal: '0 .t/ D 's 1.t/.
0.6
=0.5
0.5
=0.7
0.4
=1
() s / n
0.3
=2
0.2
0.1
=5
0.1
10
12
=n t
Figure 14.9. Plots of . / as a function of D !n t, for a second-order loop lter with a frequency step input signal: '0 .t/ D !s t 1.t/.
1039
1.4
1.2
=0.5
1
=0.7 =1
0.8
() r / n2
0.6
=2
0.4
=5
0.2
10
12
=n t
Figure 14.10. Plots of . / as a function of D !n t, for a second-order loop lter with a frequency ramp input signal: '0 .t/ D !r .t2 =2/ 1.t/.
3.5
2.5
BL n
1.5
0.5
1040
we note that B L has a minimum for D 0:5, and that for > 0:5 it increases as .1=2/ ; the choice of is therefore critical and represents a trade-off between the variance of the phase error and the speed of convergence. For a detailed analysis of the second and third-order loops we refer to [1].
14.3
Costas loop
In the previous section, the PLL was presented as a structure capable of performing carrier recovery for signals of the type sCh .t/ D sin.2 f 0 t C '0 / (14.57)
and, in general, for signals that contain periodic components of period n= f 0 , with n positive integer. We now discuss carrier recovery schemes for both PAM-DSB (see Appendix 7.C) and QAM (see Section 7.3) signals; these signals do not contain periodic components, but are cyclostationary and hence have periodic statistical moments. We express the generic received signal sCh .t/, of PAM-DSB or QAM type, in terms of .bb/ the complex envelope sCh .t/, that for simplicity we denote by a.t/, as sCh .t/ D Re[a.t/ e j .2 f 0 tC'0 / ] (14.58)
If in the reference carrier of the complex envelope we include also the phase '0 , (7.39) becomes equal to (14.5), and 1 .bb/ a.t/ sC .t/ D p sCh .t/ D p 2 2 The expression of sC .t/, apart from the delay "T and the phase offset e given by (7.42). The autocorrelation of sCh .t/ is given by (see also (1.304)) rsCh .t; t / D
1 2
(14.59)
j arg GCh . f 0 / ,
is
Re[ra .t; t
1 2
/ e j2 f 0 ] / e j[2 f 0 .2t
/C2'0 ]
(14.60) ]
Re[raa .t; t
from which the statistical power is obtained as MsCh .t/ D rsCh .t; t/ D
1 2
E[ ja.t/j2 ] C
1 2
(14.61)
14.3.1
PAM signals
We assume that the channel frequency response GC . f /, obtained from (14.5), is Hermitian;2 then gC .t/ and a.t/ are real valued, hence ja.t/j2 D [a.t/]2 (14.62)
In practice it is sufcient that GC . f / is Hermitian in a small interval around f D 0 (see page 32).
1041
and (14.61) becomes MsCh .t/ D 1 Ma .t/[1 C cos.4 f 0 t C 2'0 /] (14.63) 2 As a.t/ is a cyclostationary random process with period T (see Example 1.9.9 on page 69, Ma .t/ is periodic of period T ; therefore MsCh .t/ is also periodic and, assuming 1=T f 0 , which is often veried in practice, its period is equal to 1=.2 f 0 /. Suppose the signal sCh .t/ is input to a device (squarer) that computes the square of the signal. The output of the squarer has a mean value (deterministic component) equal to the statistical power of sCh .t/, given by (14.63); if the squarer is cascaded with a narrow passband lter H N . f / (see Figure 14.12), with H N .2 f 0 / D 1, then the mean value of the lter output is a sinusoidal signal with frequency 2 f 0 , phase 2'0 (in practice we need to sum also the phase introduced by the lter), and amplitude .1=2/ Ma .t/. Assuming that H N . f / completely suppresses the components at low frequencies, the output lter signal is given by y.t/ D 1 Re[.h .bb/ a 2 /.t/ e j[4 f 0 tC2'0 ] ] (14.64) N 2 This expression is obtained from the scheme of Figure 14.13 in which the product of two generic passband signals, y1 .t/ and y2 .t/, is expressed as a function of their complex envelopes, and decomposed into a baseband component z 1 .t/ and a passband component z 2 .t/ centered at 2 f 0 . If H N . f / is Hermitian around the frequency 2 f 0 , then .h .bb/ a 2 /.t/ is real valued,3 N and y.t/ can be written as .h .bb/ a 2 /.t/ y.t/ D N cos.4 f 0 t C 2'0 / 2 (14.65)
Figure 14.13. Baseband and passband components of the product of two generic passband signals, y1 .t/ and y2 .t/, as a function of their complex envelopes.
.bb/
1042
Thus we have obtained a sinusoidal signal with frequency 2 f 0 , phase 2'0 , and slowly varying amplitude, function of the bandwidth of H N . f /. The carrier can be reconstructed by passing the signal y.t/ through a limiter, which eliminates the dependence on the amplitude, and then to a frequency divider that returns a sinusoidal signal with frequency and phase equal to half those of the square wave. In the case of a time-varying phase '0 .t/, the signal y.t/ can be sent to a PLL with a VCO that operates at frequency 2 f 0 , and generates a reference signal equal to vVCO .t/ D A sin.4 f 1 t C 2'1 .t// D A sin.4 f 0 t C 2'0 .t// O (14.66)
The signal vVCO must then be sent to a frequency divider to obtain the desired carrier. The block diagram of this structure is illustrated in Figure 14.14. Observe that the passband lter H N . f / is substituted by the lowpass lter H L P F . f /, with H L P F .0/ D 1, inserted in the feedback loop of the PLL; this structure is called the squarer/PLL. An alternative tracking structure to the squarer/PLL, called a Costas loop, is shown in Figure 14.15. In a Costas loop the signal e.t/ is obtained by multiplying the I and Q components of the signal r.t/; the VCO directly operates at frequency f 0 , thus eliminating the frequency divider, and generates the reconstructed carrier p vVCO .t/ D 2A cos.2 f 0 t C '0 .t// O (14.67) By the equivalences of Figure 14.13 we nd that the input of the loop lter is identical to that of the squarer/PLL, and is given by e.t/ D where is the phase error (14.14). A 2 a .t/ sin[2 .t/] 4 (14.68)
14.3.2
QAM signals
The schemes of Figures 14.12 and 14.14 cannot be directly applied to QAM systems. Indeed, the symmetry of the constellation of a QAM system usually leads to E[a 2 .t/] D 0 (see (1.403)); therefore the periodic component in (14.61) is suppressed.
1043
We compute now the fourth power of sCh .t/; after a few steps we obtain
4 sCh .t/ D 1 8
ja.t/j4
4 Filtering sCh by a passband lter centered at 4 f 0 (see Figure 14.16), eventually followed by a PLL in the case of a time-varying phase, we obtain a signal y.t/ having a mean value given by
E[y.t/] D
1 4
fE[a 4 .t/] I
2 Ma I g cos.8 f 0 t C 4'0 /
(14.70)
which is a periodic signal with period 1=.4 f 0 / and phase 4'0 . In the case of M-PSK signals, there exists a variant of the Costas loop called extended Costas loop; the scheme for a QPSK system is illustrated in Figure 14.17. In the presence of additive noise, a passband lter centered at f 0 is placed in front of all schemes to limit the noise without distorting the desired signal. For the performance analysis we refer to [2, 3], in which similar conclusions to those described in Section 14.2 for the PLL are reached.
1044
14.4
With reference to the signal model (14.7), once the carrier has been recovered by one of the two methods described in the previous sections, after demodulation the three parameters .; "; / need to be estimated. In this section the optimum receiver is obtained using the ML criterion discussed in Chapter 6 (see also (2.199)). The next two sections synthetically describe various estimation methods given in [4, 5]. From (14.7) the received signal rC .t/ can be expressed as rC .t/ D e j .
tC /
(14.71)
ak qC .t
kT
"T /
(14.72)
We express (14.71) using the vector notation, that is rDsCw with the following assumptions: 1. the phase offset is equal to z, 2. the time shift "T is equal to eT ,
4
(14.73)
1045
is equal to o,
4. the transmitted data sequence a is equal to ; then the probability density function of r is given by (6.16), that is 1 2 jj sjj prj;"; ;a . j z; e; o; / D K exp N0
(14.74)
The quantities jjjj2 and jjsjj2 are constants5 [3]; therefore (14.74) is proportional to the likelihood 2 L;"; ;a .z; e; o; / D exp Re[h; si] (14.75) N0 Referring to the transmission of K symbols or to a sufciently large observation interval TK D K T , (14.75) can be written as Z 2 .t/ sC .t/ e j .otCz/ dt (14.76) L;"; ;a .z; e; o; / D exp Re N 0 TK Inserting the expression (14.72) of sC .t/, limited to the transmission of K symbols, in (14.76) and interchanging the operations of summation and integration we obtain ( ) Z X 2 K 1 jz jot Re k e .t/ e qC .t kT eT / dt L;"; ;a .z; e; o; / D exp N0 kD0 TK (14.77) 6 We introduce the matched lter
g M .t/ D qC . t/
(14.78)
and assume that the pulse .qC g M /.t/ is a Nyquist pulse; therefore there exists a suitable sampling phase for which ISI is avoided. Finally, if we denote the integral in (14.77) by x.kT C eT; o/, that is x.kT C eT; o/ D .rC . / e (14.77) becomes ( L;";
;a .z; e; o; / jo
g M . //.t/jtDkT CeT )
jz
(14.79)
D exp
(14.80)
Let us now suppose that the optimum values of z; e; o that maximize (14.80), i.e. the estimates of z; e; o have been determined in some manner.
5
Here we are not interested in the detection of a, as in the formulation of Section 8.10, but rather in the estimate of the parameters , ", and ; in this case, if the observation is sufciently long we can assume that jjsjj2 is invariant with respect to the different parameters. 6 In this formulation, the lter g is anticausal; in practice a delay equal to the duration of q must be taken M C into account.
1046
rC (t) e^ jt
rD (t)
gM (t)
x(t)
^ ak
The structure of the optimum receiver derived from (14.80) is illustrated in Figure 14.18. The signal rC .t/ is multiplied by expf j O tg, where O is an estimate of , to remove the frequency offset, then ltered by the matched lter g M .t/ and sampled at the sampling instants kT C " T , where " is an estimate of ". The samples x.kT C " T; O / are then O O O O multiplied by expf j g to remove the phase offset. Finally, the data detector decides on the symbol ak that, in the absence of ISI, maximizes the k-th term of the summation in O O O (14.80) evaluated for .z; e; o/ D . ; " ; O /:
O ak D arg max Re[k x.kT C " T; O / e O k O j
(14.81)
The digital version of the scheme of Figure 14.18 is illustrated in Figure 14.19; it uses an anti-aliasing lter and a sampler with period Tc such that (recall the sampling theorem on page 30) 1 max BC 2Tc 2 (14.82)
Observation 14.1 To simplify the implementation of the digital receiver, the ratio F0 D T =Tc is chosen as an integer; in this case, for F0 D 4 or 8, the interpolator lter may be omitted and the timing after the matched lter has a precision of Tc D T =F0 . This approach is usually adopted in radio systems for the transmission of packet data. To conclude this section, we briey discuss the algorithms for timing and carrier phase recovery.
Timing recovery
Ideally, at the output of the anti-aliasing lter the received signal should be sampled at the instants t D kT C "T ; however, there are two problems: 1. the value of " is not known; 2. the clock at the receiver allows sampling at multiples of Tc , not at multiples of T , and the ratio T =Tc is not necessarily a rational number. Therefore time synchronization methods are usually composed of two basic functions. 1) Timing estimate. The rst function gives an estimate " of ". O
Figure 14.19. Digital receiver for QAM systems. [From Meyr, Moeneclaey, and Fechtel (1998). Reproduced by permission of Wiley.]
1047
1048
2) Interpolation and decimation. The sampling instants t D kT C " T can be written as O T T kT C " T D k O C" O (14.83) Tc Tc Tc The expression within brackets admits the following decomposition: T T T T k D k C k C" O C" O Tc Tc Tc Tc D m k C k Given a real number a, bac denotes the largest integer smaller than or equal to a (oor), and D a bac is the fractionary part that we denote as [a] F . Suppose now that the estimate of " is time varying; we denote it by "k . Consider the .k C 1/-th sampling instant, O expressed as .k C 1/ T C "kC1 T O By adding and subtracting "k T , (14.85) can be rewritten as follows: O O " .k C 1/ T C "kC1 T D kT C "k T C T C .O kC1 O Substituting kT C "k T with mk Tc C k Tc , we obtain O .k C 1/ T C "kC1 T D mk Tc C k Tc C T C .O kC1 O " T D mk C k C .1 C .O kC1 " Tc "k / T O "k // Tc O "k / T O (14.86) (14.85)
(14.84)
(14.87)
Recalling that mk is a positive integer, and k is real valued and belongs to the interval [0,1), from (14.87) the following recursive expressions for mk and k are obtained: T " O mkC1 D mk C k C .1 C .O kC1 "k // Tc (14.88) T " O kC1 D k C .1 C .O kC1 "k // Tc F O " O The quantity saw.O kC1 "k / is often substituted for .O kC1 "k /, where saw.x/ is the " saw-tooth function illustrated in Figure 14.20. Thus, the difference between two successive estimates belongs to the interval [ 1=2; 1=2]; this choice reduces the effects that a wrong estimate of " would have on the value of the pair .mk ; k /. Figure 14.21 illustrates the graphic representation of (14.84) in the ideal case " D ". O The transmitter time scale, dened by multiples of T , is shifted by a constant quantity equal to "T . The receiver time scale is dened by multiples of Tc . The fact that the ratio T /Tc may be a non-rational number has two consequences: rst, the time shift k Tc is time varying even if "T is a constant; second, the instants mk Tc form a non-uniform subset of the receiver time axis, such that on average the considered samples are separated by an interval T .
1049
a)
(k-1)T+ T
kT+ T
(k+1)T+ T
(k+2)T+ T
(k+3)T+ T k+3 Tc
(k+4)T+ T k+4 Tc
k-1 Tc
k Ts
k+1 Tc
k+2 Tc
b)
m k-1 Tc m k Tc m k+1 Ts m k+2 Tc m k+3 Tc m k+4 Tc lTc
Figure 14.21. (a) Transmitter time scale; (b) receiver time scale with Tc < T, for the ideal case " D ". [From Meyr, Moeneclaey, and Fechtel (1998). Reproduced by permission of Wiley.]
With reference to Figure 14.19, to obtain the samples of the signal x.t/ at the instants .mk C k / Tc we can proceed as follows: a) implement a digital interpolator lter that provides samples of the received signal at the instants .n C k / Tc , starting from samples at the instants nTc (see Section 1.A.5); b) implement a downsampler that yields samples at the instants .mk Ck / Tc D kT C " T . O With regard to the digital interpolator lter, consider a signal r D .t/ with bandwidth Br D 1=.2Tc /. From the sampling theorem, the signal r D .t/ can be reconstructed from its samples r D .i Tc / using the relation C1 X t i Tc (14.89) r D .t/ D r D .i Tc / sinc Tc iD 1 This expression is valid for all t; in particular it is valid for t D t1 C k Tc , thus yielding the signal r D .t1 C k Tc /. Sampling this signal at t1 D nTc we obtain r D .nTc C k Tc / D
C1 X iD 1
r D .i Tc / sinc.n C k
i/
(14.90)
1050
Observe that the second member of (14.90) is a discrete-time convolution; in fact, introducing the interpolator lter with impulse response h I and parameter k , h I .i Tc ; k / D sinc.i C k / (14.90) can be rewritten as r D .nTc C k Tc / D [r D .i Tc / h I .i Tc ; k /] .nTc / (14.92) iD 1; : : : ; C1 (14.91)
In other words, to obtain from samples of r D .t/ at instants nTc the samples at nTc C k Tc we can use a lter with impulse response h I .i Tc ; k /.7 With regard to the cascade of the matched lter g M .i Tc / and the decimator at instants mk Tc , we point out that a more efcient solution is to implement a lter with input at instants nTc that generates output samples only at instants mk Tc . We conclude this section by recalling that, if after the matched lter g M , or directly in place of g M , there is an equalizer lter c with input signal having sampling period equal to Tc , the function of the lter h I is performed by the lter c itself (see Section 8.4).
In practice, the lter impulse response h I .; k / must have a nite number N of coefcients. The choice of N depends on the ratio T =Tc and on the desired precision; for example, for T =Tc D 2 and a normalized MSE, given by Z 1 Z 1=.2T / J D 2T je j2 f H I . f ; /j2 d f d (14.93)
0 1=.2T /
where H I . f ; / D
N =2 X iD .N =2/C1
h I .i Tc ; / e j2 f i Tc
(14.94)
equal to 50 dB, it turns out N ' 5. Of course, more efcient interpolator lters than that dened by (14.91) can be utilized.
1051
2) Phase rotation. O a) The samples x.kT C " T; O / are multiplied by the complex signal exp. j .kT // O (see Figure 14.19); a possible residual frequency offset 1 can be corrected by a d O O time-varying phase given by .kT / D C kT 1 . b) The samples x.kT C " T; O / e O the true values of .; "/.
O j
3) Frequency synchronization. A rst coarse estimate of the frequency offset needs to be performed in the analog domain. In fact, algorithms for timing and phase recovery only work in the presence of a small residual frequency offset. A second block provides a ne estimate of this residual offset that is used for frequency offset compensation.
14.5
In this section we discuss digital algorithms to estimate the time shift and the carrier phase offset under the assumption of absence of frequency offset, or D 0. Thus, the output samples of the decimator of Figure 14.19 are expressed as x.kT C " T; 0/; they will be O " simply denoted as x.kT C " T /, or in compact notation as x k .O /. O
14.5.1
ML criterion
The expression of the likelihood is obtained from (14.80) assuming o D 0, that is ( ) X 2 K 1 jz L;";a .z; e; / D exp Re[k x k .e/ e ] N0 kD0 D
K 1 Y kD0
(14.95)
14.5.2
Synchronization algorithms are obtained by the ML criterion (see (14.77) or equivalently (14.74)) averaging the probability density function prj;";a . j z; e; / with respect to parameters that do not need to be estimated. Then we obtain the following likelihood functions:
1052
prj;";a . j z; e; / P[a D ]
(14.96)
for the estimate of the phase, # Z "X prj . j z/ D prj;";a . j z; e; / P[a D ] p " .e/ de
(14.97)
(14.98)
With the exception of some special cases, the above functions cannot be computed in close form. Consequently we need to develop appropriate approximation techniques. A rst classication of synchronization algorithms is based on whether knowledge of the data sequence is available or not; in this case we distinguish two classes: 1. decision-directed (DD) or data-aided (DA); 2. non-data aided (NDA). If the data sequence is known, for example, by sending a training sequence a D 0 during the acquisition phase, we speak of data-aided algorithms. As the sequence a is known, in the sum in the expression of the likelihood function only the term for D 0 remains. For example, the joint estimate of .; "/ reduces to the maximization of the likelihood prj;";a . j z; e; 0 /, and we get O O . ; "/ D A D arg max prj;";a . j z; e; 0 /
z;e
(14.99)
O On the other hand, whenever we use the detected sequence a as if it were the true sequence O a, we speak of data-directed algorithms. If there is a high probability that a D a, again in the sum in the expression of the likelihood function only one term remains. Taking again the joint estimate of .; "/ as an example, in (14.96) the sum reduces to X O prj;";a . j z; e; / P[a D ] ' p rj;";a . j z; e; a/ (14.100)
O as P[a D a] ' 1. The joint estimate .; "/ is thus given by O O O . ; "/ D D D arg max prj;";a . j z; e; a/
z;e
(14.101)
Non-data aided algorithms apply instead, in an exact or approximate fashion, the averaging operation with respect to the data sequence. A second classication of synchronization algorithms is made based on the synchronization parameters that must be eliminated; then we have four cases:
1053
1. DD & D": data and timing directed, prj . j z/ D p rj;";a . j z; " ; a/ O O 2. DD, timing independent, prj . j z/ D Z O prj;";a . j z; e; a/ p" .e/ de (14.103) (14.102)
3. DD & D : data and phase directed, O O prj" . j e/ D prj;";a . j ; e; a/ 4. DD, phase independent or non-coherent (NC), Z O prj;";a . j z; e; a/ p .z/ dz prj" . j e/ D (14.104)
(14.105)
A further classication is based on the method for obtaining the timing and phase estimates: 1. feedforward (FF). The FF algorithms directly estimate the parameters .; "/ without using signals that are modied by the estimates; this implies using signals before the interpolator lter for timing recovery and before the phase rotator (see Figure 14.19) for carrier phase recovery. 2. feedback (FB). The FB algorithms estimate the parameters .; "/ using also signals that are modied by the estimates; in particular they yield an estimate of the errors O e D and e" D " ", which are then used to control the interpolator lter and O the phase rotator, respectively. In general, feedback structures are able to track slow changes of parameters. Next, we give a brief description of FB estimators, with emphasis on the fundamental blocks and on inputoutput relations.
Feedback estimators
In Figure 14.22 the block diagrams of a FB phase (FB ) estimator and of a FB timing (FB") estimator are illustrated. These schemes can be easily extended to the case of a FB frequency offset estimator. The two schemes only differ in the rst block. In the case of the FB estimator, the rst block is a phase rotator that, given the input signal s.kT / O O (x.kT C " T / in Figure 14.19), yields s.kT / exp. j k /, where k is the estimate of at O instant kT ; in the case of the FB" estimator, it is an interpolator lter that, given the input O O signal s.kT / (r D .nTc / in Figure 14.19), returns s.kT C "k T / (r D .kT C " T / in Figure 14.19), where "k is the estimate of " at instant kT . O We analyze only the FB estimator, as the FB" estimator is similar. The error detector block is the fundamental block and has the function of generating a signal e.kT /, called error signal, whose mean value is written as E[e.kT /] D g . O k / (14.106)
1054
s(kT) a)
s(kT) e-j ^ k
error detector
loop filter
e(kT)
u(kT)
F(z)
s(kT)
interpolator ^ k NCO
s(kT+^ kT)
b)
u(kT)
F(z)
where g ./ is an odd function. Then the error signal e.kT / admits the following decomposition e.kT / D g . O O k / C .kT ; ; k / (14.107)
where ./ is a disturbance term called loop noise; the loop lter F.z/ is a lowpass lter that has two tasks: it regulates the speed of convergence of the estimator and mitigates the effects of the loop noise. The loop lter output, u.kT /, is input to the numerically controlled oscillator (NCO), that updates the phase estimate according to the following recursive relation: O O kC1 D k C u.kT / (14.108)
where denotes the NCO gain. In most FB estimators the error signal e.kT / is obtained as the derivative of the likelihood, or of its logarithm, with respect to the parameter to be estimated, evaluated using O O the most recent estimates k , "k . In particular, we have the two cases @ Ok ; e; / e.kT / / L;";a . @e eDO k " (14.109) @ L;";a .z; " k ; / e.kT / / O O @z zD k We note that in FB estimators the vector a represents the transmitted symbols from instant O O 0 up to the instant corresponding to the estimates k , "k .
1055
+T x(t) kT+ ^ k T
G L [ .]
+ -
G L [ .]
loop filter
^ k NCO u(kT)
F(z)
e(kT)
Early-late estimators
Early-late estimators constitute a subclass of FB estimators, where the error signal is computed according to (14.109), and the derivative operation is approximated by a nite difference [3], i.e., given a signal p.t/, its derivative is computed as follows: p.t C / p.t d p.t/ ' dt 2 / (14.110)
where is a positive real number. Consider, for example, a DD & D timing estimator and denote by G L [] the function that, given the input signal x k .O k /, yields the likelihood L" .e/ evaluated at e D "k . From " O (14.110) the error signal is given by e.kT / / G L [x k .O k C /] " G L [x k .O k " /] (14.111)
The block diagram of Figure 14.22b is modied into that of Figure 14.23. Observe that in the upper branch the signal x k .O k / is anticipated of T , while in the lower branch it is " delayed by T : hence the name of early-late estimator.
14.5.3
Timing estimators
Non-data aided
The estimate of " can be obtained from (14.95) by eliminating the dependence on the parameters and z that are not estimated. To remove this dependence, we take the expectation with respect to a; assuming i.i.d. symbols we obtain the likelihood L;" .z; e/ D
K 1 Y kD0
E ak
jz
(14.112)
1056
For an M-PSK signal, with M > 2, we approximate ak as e j'k , where 'k is a uniform r.v. on . ; ]; then (14.112) becomes L;" .z; e/ D
K 1 Z C Y kD0
2 exp Re[e N0
jvk
x k .e/ e
jz
dvk ] 2
(14.113)
If we use the denition of the Bessel function (4.216), (14.113) is independent of the phase and we obtain K 1 Y jx k .e/j L" .e/ D (14.114) I0 N0 =2 kD0 On the other hand, if we take the expectation of (14.95) only with respect to the phase we obtain K 1 Y jx k .e/ j k (14.115) L";a .e; / D I0 N0 =2 kD0 We observe that, for M-PSK, L";a .e; / D L" .e/, as jk j is a constant, while this does not occur for M-QAM. To obtain estimates from the two likelihood functions just obtained, if the signal-to-noise ratio 0 is sufciently high, we utilize the fact that I0 ./ can be approximated as I0 . / ' 1 C
2
for j j 1
(14.116)
Taking the logarithm of the likelihood and eliminating non-relevant terms, we obtain the following NDA estimator and DA estimator. NDA: " D arg max lnfL" .e/g O
e
K 1 X kD0
jx k .e/j2
(14.117)
DA:
K 1 X kD0
jx k .e/j2 jk j2
(14.118)
On the other hand, if 0 1, (14.112) can be approximated using a power series expansion of the exponential function. Taking the logarithm of (14.112), using the hypothesis of i.i.d. symbols, with E[an ] D 0, and eliminating non-relevant terms, we obtain the following log-likelihood: " # K 1 K 1 X X 2 `;" .z; e/ D E[ja n j2 ] jx k .e/j2 C Re E[an ] .x k .e//2 e j2z (14.119)
kD0 kD0
1057
Averaging with respect to we obtain the following phase independent log-likelihood: `" .e/ D
K 1 X kD0
jx k .e/j2
(14.120)
which yields the same NDA estimator as (14.117). 2 For a modulation technique characterized by E[an ] 6D 0, (14.119) may be used to obtain an NDA joint estimate of phase and timing. In fact, for a phase estimate given by ( ) K 1 X 1 2 2 O arg E[an ] .x k .e// D (14.121) 2 kD0 the second term of (14.119) is maximized. Substitution of (14.121) in (14.119) yields a new estimate " given by O K 1 K 1 X X 2 2 2 2 (14.122) jx k .e/j C E[an ] x k .e/ " D arg max E[jan j ] O e kD0 kD0 The block diagram of the joint estimator is shown in Figure 14.24, where P values of the time shift ", ".m/ , m D 1; : : : ; P, equally spaced in [ 1=2; 1=2] are considered; usually the resolution obtained with P D 8 or 10 is sufcient. For each time shift ".m/ , the log-likelihood (14.122) is computed and the value of ".m/ associated with the largest value of the log-likelihood is selected as the timing estimate. Furthermore, we observe that in the generic branch m, ltering by the matched lter g M .i Tc C ".m/ T / and sampling at the instants kT can be implemented by the cascade of an interpolator lter h I .i Tc ; .m/ / (where .m/ depends on ".m/ ) and a lter g M .i Tc /, followed by a decimator that provides samples at the instants mk Tc , as illustrated in Figure 14.19 and described in Section 14.4.
jx.kT C eT /j2
(14.123)
Now, as x is a QAM signal, the process jx.kT C eT /j2 is approximately cyclostationary in e of period 1 (see Section 7.2). We introduce the following Fourier series representation jx.kT C eT /j2 D
C1 X iD 1
ci e j2i e
.k/
(14.124)
where the coefcients fci.k/ g are random variables given by Z 1 jx.kT C eT /j2 e j2i e de ci.k/ D
0
(14.125)
1058
Figure 14.24. NDA joint timing and phase (for E[a2 ] 6D 0) estimator. [From Meyr, Moeneclaey, and Fechtel (1998). Reproduced by n permission of Wiley.]
1059
Now (14.123) is equal to the average of the cyclostationary process jx.kT C eT /j2 in the interval [ L; L]; dening ci D
L X kD L
ci.k/
(14.126)
it results [4] that only c0 and c1 have non-zero mean, and (14.123) can be written as X `" .e/ D c0 C 2Re[c1 e j2 e ] C 2Re[ci e j2i e ] (14.127)
disturbance with zero mean for each value of e
ji j2
{z
As c0 and jc1 j are independent of e, the maximum of `" .e/ yields "D O 1 arg c1 2 (14.128)
However, the coefcient c1 is obtained by integration, which in general is hard to implement in the digital domain; on the other hand, if the bandwidth of jx.lT /j2 satises the relation Bjxj2 D 1 1 .1 C / < T 2Tc (14.129)
where is the roll-off factor of the matched lter, then c1 can be computed by DFT. Let F0 D T =Tc , then we obtain c1 D
L X kD L
"
j .2=F0 /l
# (14.130)
A simple implementation of the estimator is possible for F0 D 4; in fact, in this case no multiplications are needed; as e j .2=4/l D . j/l , (14.130) simplies into " # L 3 X 1 X c1 D jx.[4k C l] Tc /j2 . j/l (14.131) 4 lD0 kD L Figure 14.25 illustrates the implementation of the estimator for F0 D 4.
1060
( .) k=-L
2
Im(c 1 )
| x [ (4k-1)T
(4kTc )
c]|
| z [(4k-3)T c ] |
1 arg (c1 ) 2
| x(nTc )| 2
Tc
Tc
Tc
(4kTc )
| x(4kTc ) | 2
| x [(4k-2)T
c ]| 2
( .) k=-L
Re(c1 )
Figure 14.25. NDA timing estimator via spectral estimation for the case F0 D 4. [From Meyr, Moeneclaey, and Fechtel (1998). Reproduced by permission of Wiley.]
Figure 14.26. Phase independent DA (DD) timing estimator. [From Meyr, Moeneclaey, and Fechtel (1998). Reproduced by permission of Wiley.]
from which we immediately derive the estimate " D arg max L" .e/ O
e
(14.133)
The block diagram of the estimator is shown in Figure 14.26; note that this algorithm can only be used in the case phase recovery is carried out before timing recovery. For a joint phase and timing estimator, from (14.95) we get ( L;" .z; e/ D exp " K 1 X 2 Re ak x k .e/ e O N0 kD0 #)
jz
(14.134)
1061
Dening r.e/ D the estimation algorithm becomes O O .; " / D arg max Re[r.e/ e
z;e jz K 1 X kD0
ak x k .e/ O
(14.135)
] (14.136) ]
j .z arg.r.e///
The two-variable search of the maximum reduces to a single-variable search; as a matter of fact, once the value of e that maximizes jr.e/j is obtained, which is independent of z, the second term Re[e
j .z arg.r.e///
(14.137)
is maximized by z D arg.r.e//. Therefore the joint estimation algorithm is given by K 1 X ak x k .e/ O " D arg max jr.e/j D arg max O e e kD0 (14.138) O D arg r.O / D arg "
K 1 X kD0
ak x k .O / O "
Figure 14.27 illustrates the implementation of this second estimator; note that this scheme is a particular case of (7.269). For both estimators, estimation of the synchronization parameters is carried out every K samples, according to the assumption of slow parameter variations made at the beginning of the section.
^ a k* x k ( (1)) kT r AA (t) nTc g (iTc + (P) ) M kT xk ( (P) )
arg max r ( )
g (iTc + (1) )
M
(.)
k
r ( 1 )
(.)
k
arg r ( ^ )
r ( P )
Figure 14.27. DA (DD) joint phase and timing estimator. [From Meyr, Moeneclaey, and Fechtel (1998). Reproduced by permission of Wiley.]
1062
Observation 14.2 In the case the channel is not known, to implement the matched lter g M we need to estimate the overall impulse response qC ; then the estimation of qC , for example, by one of the methods presented in Appendix 3.A, and of timing can be performed jointly. Let F0 D T =Tc and Q 0 D T =TQ be integers, with Q 0 F0 . From the signal fr A A .q TQ /g, obtained by oversampling r A A .t/ or by interpolation of fr A A .nTc /g, and the knowledge of the training sequence fak g, k D 0; : : : ; L T S 1, the estimate of qC with sampling period TQ , or equivalently the estimate of its Q 0 =F0 polyphase components with sampling period Tc (see Observation 8.5), is made. Limiting the estimate to the more signicant consecutive samples around the peak, the determination of the timing phase with precision TQ coincides with the selection of the polyphase component with the largest energy among the Q 0 =F0 polyphase components. This determines the optimum lter g M with sampling period Tc . Typically for radio systems F0 D 2, and Q 0 D 4 or 8.
In practice, if T =Tc 2 it is simpler to implement a differentiator lter by a nite difference lter having a symmetric impulse response given by h d .i Tc / D 1 .i C1 2Tc i
1/
(14.142)
Figure 14.28 illustrates the block diagram of the estimator, where the compact notation x.t/ is used in place of .dx.t/=dt/; moreover, based on the analysis of Section 14.5.2, if P u.kT / is the loop lter output, the estimate of " is given by O "kC1 D "k C " u.kT / O (14.143)
O where " is a suitable constant. Applying (14.88) to the value of "kC1 , we obtain the values of kC1 and mkC1 .
Figure 14.28. DD & D -FB timing estimator. [From Meyr, Moeneclaey, and Fechtel (1998). Reproduced by permission of Wiley.]
1063
1064
Observe that, under the assumptions of Section 14.4, q R .t/ is a Nyquist pulse; moreover, we assume that in the absence of channel distortion, q R .t/ is an even function. Note that the signal (14.144) is an odd function of the estimation error e" for e" 2 . 1; 1/, whereas the signal (14.145) is an odd function of e" only around e" D 0. Under lock conditions, i.e. for e" ! 0, the two versions of the algorithm exhibit a similar behavior. However, the type A algorithm gives better results than the type B algorithm in transient conditions because the mean value of the error signal for the type B algorithm is not symmetric. Moreover, the type A algorithm results effective also in the presence of signal distortion. The error signal for the type A algorithm is chosen equal to e.kT / D Re[[ak O " 1 x k .O / ak x k O " 1 .O /] e
O j
(14.146)
where is a suitable constant whose value is discussed below. Assuming that ak 1 D ak 1 , O O ak D ak , and D , from (14.71) and (14.79) for o D D 0 (14.146) can be written as O (" C1 X e.kT / D Re ak 1 ai q R .kT C e" T i T /
iD 1 ak C1 X iD 1
C ak
Q 1 wk
a k wk Q
where wk is the decimated noise signal at the matched lter output. We dene Q qm .e" / D q R .mT C e" T / then with a suitable change of variables (14.147) becomes (" # C1 C1 X X e.kT / D Re ak 1 ak m qm .e" / ak ak 1 m qm .e" / C ak wk Q
mD 1 mD 1
(14.148)
)
a k wk 1 Q
(14.149)
1065
Taking the mean value of e.kT / we obtain E[e.kT /] D Ref.E[jak j2 ] D Ref.E[jak j2 ] For D 1 2.E[jak j2 ] jma j2 / (14.151) jma j2 /[q1 .e" / q
1 .e" /]g
(14.150) T /]g
q R .e" T
we obtain (14.144). Similarly, in the case of the type B algorithm the error signal assumes the expression e.kT / D 1 .E[jak j2 ] jma j2 / [ak O
1
ma ] x k .O / e "
O j
(14.152)
Figure 14.29 illustrates the block diagram of the direct section of the type A estimator. The constant is included in the loop lter and is not explicitly shown.
2Re[x.kT C eT / x .kT C eT /] P
xk ( ^ )
^ ak ( . )* ^ ak* T
^ e -j
+ -
Re[ .]
e(kT)
T
Figure 14.29. Mueller & Muller type A timing estimator.
1066
If we assume that the sum is carried out by the loop lter, the error signal is given by P O e.kT / D Re[x.kT C "k T / x .kT C "k T /] O (14.154)
14.5.4
Phasor estimators
(14.156)
(14.157)
where wk represents the decimator output noise, and w M;k denotes the overall disturbance. Q M As ak D .e j2l=M / M D 1, (14.157) becomes
M " x k .O / D e j M C w M;k From (14.95), we substitute .x k .O // M for k x k .O / obtaining the likelihood " " ( " #) K 1 X 2 M j zM L .z/ D exp Re .x k .O // e " N0 kD0
(14.158)
(14.159)
K 1 X kD0
# .x k .O // M " (14.160)
^ a k* xk ( ^ ) k=0
e j
1067
xk ( ^ )
( .) M
arg ( xk ( ^ ) ) M
k=0
K-1
j^ e M
We note that raising x k .O / to the M-th power causes a phase ambiguity equal to a multiple " O O of .2 /=M; in fact, if is a solution to (14.160), also . C 2l=M/ for l D 0; : : : ; M 1, are solutions. This ambiguity can be removed, for example, by differential encoding (see Section 6.5.2). The estimator block diagram is illustrated in Figure 14.31.
(14.161)
" Observe that, in the absence of noise, x k .O / D ak e j , and (14.161) becomes e.kT / D jak j2 sin. O
pk a) ek uk
O k /
^ p k+1 ^ pk
(14.162)
F(z)
loop filter
z -1
NCO
~ ^ xk ( )=ak e j+w k b) ( .) * 1/ . ^ pk
^ ak
( .) * ^* ak
pk
PHLL
Figure 14.32. (a) PHLL; (b) DD & D"-FB phasor estimator. [From Meyr, Moeneclaey, and Fechtel (1998). Reproduced by permission of Wiley.]
1068
Hence, we can use a digital version of the PLL to implement the estimator. However, the error signal (14.161) introduces a phase ambiguity; in fact it assumes the same value if we O O substitute (k ) for k . An alternative to the digital PLL is given by the phasor-locked loop (PHLL), that provides an estimate of the phasor e j , rather than the estimate of , thus eliminating the ambiguity. The block diagram of the PHLL is illustrated in Figure 14.32a; it is a feedback structure O with the phasor pk D e jk as input and the estimate pk D e j k as output. The error signal O O ek is obtained by subtracting the estimate pk from pk ; then ek is input to the loop lter F.z/ that yields the signal u k , which is used to update the phasor estimate according to the recursive relation pkC1 D pk C . f u/.k/ O O (14.163)
Figure 14.32b illustrates the block diagram of a DD & D" phasor estimator that implements the PHLL. Observe that the input phasor pk is obtained by multiplying x k .O / by ak " O to remove the dependence on the data; the dashed block normalizes the estimate pk in the O QAM case.
14.6
As mentioned in Section 14.4, phase and timing estimation algorithms work correctly only if the frequency offset is small. Therefore the frequency offset must be compensated before the estimate of the other two synchronization parameters takes place. Hence the algorithms that we will present are mainly NDA and non-clock-aided (NCA); timingdirected algorithms are possible only in the case the frequency offset has a magnitude much smaller than 1=T . In Figure 14.33 we redraw part of the digital receiver scheme of Figure 14.19; observe that the position of the matched lter is interchanged with that of the interpolator lter. In this scheme, the samples r A A .nTc / are multiplied by exp. j O nTc / to remove the frequency offset. In [4] it is shown that, whenever satises the condition T (14.164) 2 0:15
Figure 14.33. Receiver of Figure 14.19 with interpolator and matched lter interchanged. [From Meyr, Moeneclaey, and Fechtel (1998). Reproduced by permission of Wiley.]
1069
x k .e/
(14.165)
Then the likelihood (14.80) can be written as ( X 2 K 1 Re[k x k .e/ e L;"; ;a .z; e; o; / D exp N0 kD0
)
jokT
jz
(14.166)
Therefore in the schemes of Figure 14.19 and Figure 14.33 the frequency translator may be moved after the decimator, together with the phase rotator.
14.6.1
Non-data aided
Suppose the receiver operates with a low signal-to-noise ratio 0; similarly to (14.123) the log-likelihood for the joint estimate of ."; / in the observation interval [ L T; L T ] is given by `"; .e; o/ D
L X kD L
(14.167)
By expanding `"; .e; o/ in Fourier series and using the notation introduced in the previous section we obtain X 2Re[ci e j2i e ] (14.168) `"; .e; o/ D c0 C 2Re[c1 e j2 e ] C |
ji j2 disturbance
{z
Now the mean value of c0 , E[c0 ], depends on o, but is independent of e and furthermore is maximized for o D O ; hence O D arg max c0
o
(14.169)
As we did for the derivation of (14.131), starting with (14.169) and assuming the ratio F0 D T =Tc is an integer, we obtain the following joint estimate of . ; "/ [4]: O D arg max
o LX 1 F0 nD L F0
" D arg O
LX 1 F0 nD L F0
jx.nTc ; o/j2 e
j2 n=F0
1070
e -j
(1)nTc
g (iTc )
M
( .) 2
arg max
e -j
(M) nTc
r AA (t) nTc
^ x(nTc, )
to timing and/or phase estimator
g (iTc )
M
( .) 2
Figure 14.34. NDA frequency offset estimator. [From Meyr, Moeneclaey, and Fechtel (1998). Reproduced by permission of Wiley.]
The implementation of the estimator is illustrated in Figure 14.34; observe that the signal x.nTc ; o/ can be rewritten as X x.nTc ; o/ D r A A .i Tc / e joiTc g M .nTc i Tc /
i
De De
jonTc
X
i
r A A .i Tc / e
jo.i n/Tc
g M .nTc i Tc ; o/
i Tc /
(14.171)
jonTc
X
i
r A A .i Tc / g M .nTc
. pb/
where the expression of the lter g M .i Tc ; o/ D g M .i Tc / e joiTc depends on the offset o. Dening xo .nTc / D X
i . pb/
(14.172)
r A A .i Tc / g M .nTc
. pb/
i Tc ; o/
(14.173)
we note that jx.nTc ; o/j D jxo .nTc /j, and hence in the m-th branch of Figure 14.34 the cascade of the frequency translator and the lter can be substituted with a simple lter with . pb/ impulse response g M .i Tc ; .m/ /.
1071
g M .nTc
i Tc /
(14.175)
We dene the frequency matched lter as g F M .i Tc / D .jiTc / g M .i Tc / Observe now that, if the signal r D .nTc / D r A A .nTc / e then from (14.175) the output is given by x F M .nTc / D g F M r D .nTc / D from which we obtain @ x.nTc ; o/ D x F M .nTc / @o jnTc x.nTc ; o/ (14.178)
jonTc
(14.177)
(14.179)
The block diagram of the resultant estimator is shown in Figure 14.35. The loop lter output u.kTc / is sent to the NCO that yields the frequency offset estimate according to the recursive equation O nC1 Tc D O n Tc C u.nTc / where is the NCO gain. (14.180)
(14.181)
Proceeding as in the previous section, we get the block diagram illustrated in Figure 14.36.
1072
matched filter
gM (iTc )
frequency matched filter
^ x(nT ,n ) c
FM
(iTc )
^ xFM (nTc ,n )
( .) *
2Re{ .}
NCO
u(nTc )
loop filter
F(z)
e(nT ) c
Figure 14.35. NDA-ND"-FB frequency offset estimator. [From Meyr, Moeneclaey, and Fechtel (1998). Reproduced by permission of Wiley.]
k
matched filter
m k Ts
decimator
g (iTc ) M
frequency matched filter
^ x(nTc ,n )
interpolator
h I (iTc , k )
k
^ xFM (nTc ,n )
interpolator
m k Tc
decimator
g (iTc )
FM
h I (iTc , k )
( .) *
u(kT)
F(z)
e(kT)
2Re{ .}
14.6.2
As mentioned at the beginning of Section 14.6, whenever the condition (14.164) is satised, the frequency translator can be moved after the decimator and consequently frequency offset estimation may take place after timing estimation, thus obtaining D" algorithms. The likelihood (14.166) for an observation interval [ .K 1/=2; .K 1/=2], with K odd, evaluated at e D " becomes O ( ) .KX 1/=2 2 jokT jz Re[k x k .O / e " e ] (14.182) L; ;a .z; o; / D exp N0 kD .K 1/=2
1073
jz
(14.183)
Note that the joint estimate is computed by nding the maximum of a function of one variable; in fact, dening r.o/ D (14.183) can be rewritten as O . O ; / D arg max jr.o/j Re[e
z;o j .z argfr.o/g/ .KX 1/=2 kD .K 1/=2 0;k x k .O / e " jokT
(14.184)
(14.185)
The maximum (14.183) is obtained by rst nding the value of o that maximizes jr.o/j, O D arg max jr.o/j
o
(14.186)
and then nding the value of z for which the term within brackets in (14.185) becomes real valued, O D argfr. O /g (14.187)
We now want to solve (14.186) in close form; a necessary condition to get a maximum is that the derivative of jr.o/j2 D r.o/ r .o/ with respect to o is equal to zero for o D O . Dening 2 1 K k.k C 1/ (14.188) bk D 4 we obtain O T D arg (
.KX 1/=2 kD .K 1/=2 0;kC1 0;k
bk
1074
14.7
In feedback systems, the recovery of the phase and the frequency performed by a second-order digital PLL (DPLL), given by O O kC1 D k C e .kT / C O k O kC1 D O k C
;1 e .kT / C ;2 e
where e and e are estimates of the phase error and of the frequency error, respectively, and , ;1 and ;2 are suitable constants. Typically (see Example 15.6.4 on page 1110) p ;1 ' ;2 ' . Observe that (14.191) and (14.192) form a digital version of the second-order analog PLL illustrated in Figure 14.7.
14.8
In this section we discuss early-late FB schemes, named delay-locked loops (DLLs) [7], for the timing recovery in spread spectrum systems (see Chapter 10).
14.8.1
Transmitter
In Figure 14.37 (see Figures 10.2 and 10.3) the (baseband equivalent) scheme of a spread spectrum transmitter is illustrated. The symbols fak g, generated at the modulation rate 1=T , are input to a holder, which outputs the symbols am at the chip rate 1=Tchi p . The chip N period Tchi p is given by Tchi p D T NS F (14.193)
where N S F denotes the spreading factor. Then the symbols fak g are multiplied by the spreading code fcm g and input to a lter with impulse response h T x .t/ that includes the DAC. The (baseband equivalent) transmitted signal s.t/ is expressed as s.t/ D
C1 X mD 1
am cm h T x .t N
mTchi p /
(14.194)
1075
In terms of the symbols fak g, we obtain the following alternative representation that will be used next: s.t/ D
C1 X kD 1
ak
CN k N S FXS F 1 mDk N S F
cm h T x .t
mTchi p /
(14.195)
Optimum receiver
With the same assumptions of Section 14.1 and for the transmission of K symbols, the received signal rC .t/ is expressed as rC .t/ D
K 1 X kD0
ak
CN k N S FXS F 1 mDk N S F
cm qC .t
mTchi p
"Tchi p / e j .
tC /
C wC' .t/
(14.196)
;a .z; e; o; /
( Lss D exp Z
TK
" X 2 K 1 Re k e N0 kD0
jot
jz
cm
(14.197)
. / e
jo
g M .t
/ d (14.198)
cm
x.mTchi p C eTchi p ; o/
( L;";
;a .z; e; o; /
D exp
)
jz
(14.199)
To obtain the samples y.kT; e; o/ we can proceed as follows: 1. obtain the samples y.lTchi p ; e; o/ D
l X mDl N S F C1 cm x.mTchi p C eTchi p ; o/
(14.200)
1076
^ x(mTchip + ^ Tchip , )
i=mN
* cm
^ ak
SF +1
^ ^ y(mTchip , , )
phase estimator
By (14.199) it is possible to derive the optimum digital receiver. In particular, up to the decimator that outputs samples at instants fmm Tc g the receiver is identical to that of Figure 14.19;8 then in Figure 14.38 only part of the receiver is shown.
14.8.2
Assume there is no frequency offset, i.e. in (14.199) to yield L;";a .z; e; / D exp where we use the compact notation
(14.201)
yk .e/ D y.kT; e; 0/
(14.202)
The early-late FB estimators are obtained by approximating the derivative of (14.201) with respect to e with a nite difference, and evaluating it for e D "k . O
jyk .e/j2
(14.203)
From (14.110) the derivative of `" .e/ is approximated as X 1 K 1 @`" .e/ ' [jyk .e C /j2 @e 2 kD0 jyk .e /j2 ] (14.204)
By including the constant 1=.2/ in the loop lter, and also assuming that the sum is performed by the loop lter we obtain the error signal e.kT / D jyk .O k C /j2 "
8
jyk .O k "
/j2
(14.205)
Note that now Tc ' Tchi p =2, and the sampling instants at the decimator are such that mTchi p C " Tchi p D O mm Tc C m Tc . The estimate " is updated at every symbol period T D Tchi p N S F . O
1077
The block diagram of the estimator is shown in Figure 14.39; note that the lag and the lead equal to Tchi p are implemented by interpolator lters operating at the sampling Q Q period Tc (see (14.91)) with parameter equal, respectively, to and C, where Q D .Tchi p =Tc /. The estimator is called a non-coherent digital DLL [8, 9], as the dependence of the error signal on the pair .; a/ is eliminated without computing the estimates.
(14.206)
yk .e
// ]
(14.207)
Assuming the loop lter performs the multiplication by 1= and the sum, the error signal is given by e.kT / D Re[yk .O k /.yk .O k C / " " y.O k " // ] (14.208)
The block diagram of the estimator is shown in Figure 14.40 and is called modied code tracking loop (MCTL) [10]; also in this case the estimator is non-coherent.
Re[ak yk .e/ e O
O j
(14.209)
Approximating the derivative as in (14.110) and including both the multiplicative constant and the summation in the loop lter, the error signal is given by e.kT / D Re[ak e O
O j
[yk .O k C / "
yk .O k "
/]]
(14.210)
Figure 14.41 illustrates the block diagram of the estimator, which is called coherent DLL O O [9, 10, 11], as the error signal is obtained by the estimates and a. In the three schemes of Figure 14.39, 14.40, and 14.41 the direct section of the DLL gives estimates of mm and m at every symbol period T , whereas the feedback loop may operate at the chip period Tchi p . Observe that by removing the decimation blocks the DLL is able to provide timing estimates at every chip period.
1078
1079
1080
14. Bibliography
1081
Bibliography
[1] H. Meyr and G. Ascheid, Synchronization in digital communications, vol. 1. New York: John Wiley & Sons, 1990. [2] L. E. Franks, Carrier and bit synchronization in data communicationA tutorial review, IEEE Trans. on Communications, vol. 28, pp. 11071120, Aug. 1980. [3] J. G. Proakis, Digital communications. New York: McGraw-Hill, 3rd ed., 1995. [4] H. Meyr, M. Moeneclaey, and S. A. Fechtel, Digital communication receivers. New York: John Wiley & Sons, 1998. [5] U. Mengali and A. N. DAndrea, Synchronization techniques for digital receivers. New York: Plenum Press, 1997. [6] K. H. Mueller and M. S. Muller, Timing recovery in digital synchronous data receivers, IEEE Trans. on Communications, vol. 24, pp. 516531, May 1976. [7] M. K. Simon, J. K. Omura, R. A. Scholtz, and B. K. Levitt, Spread spectrum communications handbook. New York: McGraw-Hill, 1994. [8] R. De Gaudenzi, M. Luise, and R. Viola, A digital chip timing recovery loop for bandlimited direct-sequence spread-spectrum signals, IEEE Trans. on Communications, vol. 41, pp. 17601769, Nov. 1993. [9] R. De Gaudenzi, Direct-sequence spread-spectrum chip tracking loop in the presence of unresolvable multipath components, IEEE Trans. on Vehicular Technology, vol. 48, pp. 15731583, Sept. 1999. [10] R. A. Yost and R. W. Boyd, A modied PN code tracking loop: its performance analysis and comparative evaluation, IEEE Trans. on Communications, vol. 30, pp. 1027 1036, May 1982. [11] R. De Gaudenzi and M. Luise, Decision-directed coherent delay-lock tracking loop for DS-spread spectrum signals, IEEE Trans. on Communications, vol. 39, pp. 758 765, May 1991.
Chapter 15
Self-training equalization
By the expression self-training equalization we refer to channel equalization techniques by which we obtain the initial convergence of the parameters of an adaptive equalizer without resorting to the transmission of a training sequence. Although these techniques generally achieve sub-optimum performance at convergence, they are applied in many cases of practical interest, where the transmission of known training sequences is not viable, as for example in broadcast systems, or may result in an undesirable increase of system complexity [1, 2, 3, 4, 5, 6]. In these cases it is necessary to consider self-training of the adaptive equalizer, where the terms for adjusting the coefcient parameters are obtained by processing the received signal. Usually the receiver performs self-training of the equalizer by referring to the knowledge of the input signal characteristics, for example, the probability distribution of the input symbols. The subject of self-training equalization in communications systems has received considerable attention since the publication of the article by Sato [7] in 1975; the proposed algorithms have been alternatively called self-recovering, self-adaptive, or blind.
15.1
With reference to the discrete-time equivalent scheme of Figure 7.18 shown in Figure 15.1, we consider a real-valued discrete-time system given by the cascade of a linear channel fh i g, in general with non-minimum phase transfer function1 H , and an equalizer C $ fci g; in general, the impulse responses fh i g and fci g are assumed with unlimited duration. The P overall system is given by 9 D H C $ f i g, where i D 1 1 c` h i ` . The chan`D nel input symbol sequence fak g is modeled as a sequence of i.i.d. random variables with symmetric probability density function pa ./. The channel output sequence is fx k g. In this section, additive noise introduced by the channel is ignored. We note that a non-minimum phase rational transfer function (see Denition 1.8 on page 26) can be expressed as H .z/ D H0 P1 .z/ P2 .z/ P3 .z/ (15.1)
In this chapter, to simplify notation, often the argument of the z-transform is not explicitly indicated.
1084
where H0 denotes the gain, P3 .z/ and P1 .z/ are monic polynomials with zeros inside the unit circle, and P2 .z/ is a monic polynomial with zeros outside the unit circle. We introduce the inverse functions with respect to the polynomials P1 .z/ and P2 .z/ given by P1 1 .z/ D 1=P1 .z/ and P2 1 .z/ D 1=P2 .z/, respectively. By expanding P1 1 .z/ and P2 1 .z/ in Laurent series we obtain, apart from lag factors, P1 1 .z/ D
C1 X nD0
c1;n z
(15.2) P2 .z/ D
1 0 X nD 1
c2;n z
both converging in a ring that includes the unit circle. Therefore we have ! ! C1 0 X X 1 1 n n H .z/ D P3 .z/ c1;n z c2;n z H0 nD 1 nD0
(15.3)
In general from (15.3) we note that, as the system is non-minimum phase, the inverse system cannot be described by a nite number of parameters; in other words, the reconstruction of a transmitted symbol at the equalizer output at a certain instant requires the knowledge of the entire received sequence. In practice, to obtain an implementable system, the series in (15.3) are truncated to N terms, and an approximation of H 1 with a lag equal to .N1 1/ modulation intervals is given by ! ! N 1 0 X X 1 1 .N1 1/ n n z P3 .z/ c1;n z c2;n z (15.4) H .z/ ' H0 nD0 nD .N 1/ Therefore the inverse system can only be dened apart from a lag factor. The problem of self-training equalization is formulated as follows: from the knowledge of the probability distribution of the channel input symbols fak g and from the observation of the channel output sequence fx k g, we want to nd an equalizer C such that the overall system impulse response is ISI free. Observe that, if the channel H is minimum phase,2 both H and H 1 are causal and stable; in this case the problem of channel identication, and hence the determination of
1085
the sequence fak g, can be solved by whitening (see page 143) the observation fx k g using known procedures that are based on the second-order statistical description of signals. If, as it happens in the general case, the channel H is non-minimum phase, by the second-order statistical description (1.263) it is possible to identify only the amplitude characteristic of the channel transfer function, but not its phase characteristic. In particular, if the probability density function of the input symbols is Gaussian, the output fx k g is also Gaussian and the process is completely described by a second-order analysis; therefore the above observations are valid and in general the problem of selftraining equalization cannot be solved for Gaussian processes. Note that we have referred to a system model in which the sampling frequency is equal to the symbol rate; a solution to the problem using a second-order description with reference to an oversampled model is obtained in [8]. Furthermore, we observe that, as the probability density function of the input symbols is symmetric, the sequence f ak g has the same statistical description as the sequence fak g; consequently it is not possible to distinguish the desired equalizer H 1 from the equalizer H 1 . Hence, the inverse system can only be determined apart from the sign and a lag factor. Therefore the solution to the problem of self-training equalization is given by C D H 1 , which yields the overall system 9 D I , where I denotes the identity, with the exception of a possible lag. In this chapter we will refer to the following theorem, demonstrated in [9]. Theorem 15.1 Assuming that the probability density function of the input symbols is non-Gaussian, then 9 D I if the output sample yk D
C1 X nD 1
cn x k
(15.5)
has a probability density function p yk .b/ equal to the probability density function of the symbols fak g. Therefore, using this theorem to obtain the solution, it is necessary to determine an algorithm for the adaptation of the coefcients of the equalizer C such that the probability distribution of yk converges to the distribution of ak . We introduce the cost function J D E[8.yk /] (15.6) where yk is given by (15.5) and 8 is an even, real-valued function that must be chosen so that the optimum solution, determined by Copt .z/ D arg min J
C.z/
(15.7)
is found at the points H 1 , apart from a lag factor. In most applications, minimization is done by equalizers of nite length N having coefcients at instant k given by ck D [c0;k ; c1;k ; : : : ; c N 1;k ]T (15.8)
1086
Let xk D [x k ; x k then (15.5) becomes yk D Then also (15.7) simplies into copt D arg min J
c N 1 X nD0 1 ; : : : ; x k .N 1/ ] T
(15.9)
cn;k x k
T D c k xk
(15.10)
(15.11)
If 2 is the derivative of 8, the gradient of J with respect to c is given by (see (2.18)) rc J D E[xk 2.yk /] (15.12)
For the minimization of J we use a stochastic gradient algorithm for which (see (3.40)) ckC1 D ck 2.yk / xk (15.13)
where is the adaptation gain. Note that the convergence of C to H 1 or to H 1 depends on the initial choice of the coefcients. Before tackling the problem of choosing the function 8, we consider the following problem.
ij
(15.14)
D1
(15.15)
The function V characterizes the peak amplitude of the system output signal fyk g and the constraint can be interpreted as a requirement that the solution f i g belongs to the sphere with center the origin and radius r D 1 in the parameter space f i g. Letting D [: : : ; 0 ; 1 ; : : : ]T , it results in r V D [: : : ; sgn. 0 /; sgn. 1 /; : : : ]T ; then, if max is the maximum value assumed by i , the cost function (15.14) presents stationary points for i D max . Now, taking into account the constraint (15.15), it is easily seen that the minimum of (15.14) is reached only when one element of the sequence f i g is different from zero (with a value equal to max ) and the others are all zeros. In other words, the only points of minimum are given by 9 D I , and the other stationary points correspond to saddle points. Figure 15.2 illustrates the cost function V along the unit circle for a system with two parameters.
1087
V (0, 1)
1.5
0.5
+I
0
1
0.5
0
+I
I V
0.5
1
1
I
0.5
0
0.5
Figure 15.2. Illustration of the cost function V for the system 9 $ f 0 ; 1 g and of the gradient rV projected onto the straight line tangent to the unit circle.
The minimization can also be obtained recursively, updating the parameters f i g as indicated by the direction of steepest descent3 of the cost function V. The projection of the gradient V onto the plane tangent to the unit sphere at the point is given by r V then the recursive equation yields
0 i;kC1
[.r V/T ]
C1 X `D 1
(15.16) # j
`;k j
" sgn.
0 i;kC1 0 2 `;kC1 / i;k / i;k
i;k
(15.17)
i;kC1
Dv u C1 u X t . P
`D 1
(15.18)
Note that, if the term i;k ` j `;k j is omitted in (15.17), with good approximation the direction of the steepest descent is still followed, provided that the adaptation gain is sufciently small. From (15.17), for each parameter i the updating consists of a correction toward zero of a xed value and a correction in the opposite direction of a value proportional to the
3
1088
i,0
Reduce by +
1
+
4
(1+)i,1 = i,1 i,1
6
Rescale by i,1
3
i,1
6
Reduce by +
6
Rescale by i,2
3
i,k
parameter amplitude. Assuming that the initial point is not a saddle point, by repeated iterations of the algorithm, one of the parameters i approaches the value one, while all others converge to zero, as shown in Figure 15.3. We now want to obtain the same adaptation rule for the parameters f i g using the output signal yk . We dene k as the vector of the parameters of at instant k, k D [: : : ; and ak D [: : : ; ak ; ak
1; : : : ] T 0;k ; 1;k ; : : : ] T
(15.19)
(15.20)
1089
Therefore
T yk D k ak
(15.21)
Assume that at the beginning of the adaptation process the overall system 9 satises the condition jjjj2 D 1, but it deviates signicantly from the system identity; then the equalizer output signal yk will occasionally assume positive or negative values which are much larger P in magnitude than max D max ak . The peak value of yk is given by max i j i;k j, obtained with symbols fak i g equal to max , and indicates that the distortion is too large and must be reduced. In this case a correction of a xed value towards zero is obtained using the error signal ek D yk max sgn.yk / (15.22)
If the coefcients are scaled so that the condition jj k jj2 D 1 is satised at every k, we obtain a coefcient updating algorithm that approximates algorithms (15.17) and (15.18). Obviously algorithm (15.23) cannot be directly applied, as the parameters of the overall system 9 are not available. However, observe that if the linear transformation H is non-singular, then formally C D H 1 9. Therefore the overall minima of V at the points 9 D I are mapped into overall minima at the points C D H 1 of a cost function J that is the image of V under the transformation given by H 1 , as illustrated in Figure 15.4. Furthermore, it is seen that the direction of steepest descent of V has not been modied by this transformation. Thus the updating terms for the equalizer coefcients are still given by (15.22) and (15.23), if symbols ak are replaced by the channel output samples x k . Then, a general algorithm that converges to the desired solution C D H 1 can be formulated as follows: observe the equalizer output signal fyk g and determine its peak value;
Global minima of V ( ) Global minima of J( C )
H 1 C
1090
whenever a peak value occurs, update the coefcients according to the algorithm ek D yk ckC1 D ck max sgn.yk / e k xk (15.24)
scale the coefcients so that the statistical power of the equalizer output samples is equal to the statistical power of the channel input symbols. We observe that it is not practical to implement an algorithm that requires computing the peak value of the equalizer output signal and updating the coefcients only when a peak value is observed. In the next sections we describe algorithms that allow the updating of the coefcients at every modulation interval, thus avoiding the need of computing the peak value of the signal at the equalizer output and of scaling the coefcients.
15.2
1 2
2 yk
i jyk j
(15.25)
2 E[ak ] E[jak j]
(15.26)
sgn.yk //]
(15.27)
sgn.yk /
(15.28)
which assumes the meaning of pseudo error that during self-training replaces the error used in the LMS decision directed algorithm. We recall that for the LMS algorithm the error signal is4 ek D yk ak O (15.29)
where ak is the detection of the symbol ak , obtained by a threshold detector from the sample O yk . Figure 15.5 shows the pseudo error S ;k as a function of the value of the equalizer output sample yk .
Note that in this chapter, the error signal is dened with opposite sign with respect to the previous chapters.
1091
e S,k
* gs
gS
yk
Figure 15.5. Characteristic of the pseudo error S,k as a function of the equalizer output.
Therefore the Sato algorithm for the coefcient updating of an adaptive equalizer assumes the expression ckC1 D ck S;k xk (15.30) It was proved by Benveniste, Goursat, and Ruget [9] that, if the probability density function of the output symbols fak g is sub-Gaussian,5 then the Sato cost function (15.25) admits as unique points of minimum the systems C D H 1 , apart from a possible lag; however, note that the uniqueness of the points of minimum of the Sato cost function is obtained by assuming a continuous probability distribution of input symbols. In the case of a discrete probability distribution with an alphabet A D f1; 3; : : : ; .M 1/g, the convergence properties of the Sato algorithm are not always satisfactory. Another undesirable characteristic of the Sato algorithm is that the pseudo error S;k is not equal to zero for C D H 1 , unless we consider binary transmission. In fact only the gradient of the cost function given by (15.27) is equal to zero for C D H 1 ; moreover, we nd that the variance of the pseudo error may assume non-negligible values in the neighborhood of the desired solution.
BenvenisteGoursat algorithm
To mitigate the above mentioned inconvenience, we observe that the error ek used in the LMS algorithm in the absence of noise becomes zero for C D H 1 . It is possible to combine the two error signals obtaining the pseudo error proposed by Benveniste and Goursat [10], given by G;k D 1 ek C 2 jek j S;k
5
(15.31)
A probability density function pak ./ is sub-Gaussian if it is uniform or if pak ./ D K expf g./g, where
1 dg g./ is an even function such that both g./ and d are strictly increasing in the domain [0; C1/.
1092
where 1 and 2 are constants. If the distortion level is high, jek j assumes high values and the second term of (15.31) allows convergence of the algorithm during the self-training. Near convergence, for C ' H 1 , the second term has the same order of magnitude as the rst and the pseudo error assumes small values in the neighborhood of C D H 1 . Note that an algorithm that uses the pseudo error (15.31) allows a smooth transition of the equalizer from the self-training mode to the decision-directed mode. In the case of a sudden change in channel characteristics, the equalizer is found working again in selftraining mode. Thus the transitions between the two modes occur without control on the level of distortion of the signal fyk g at the equalizer output.
Stop-and-go algorithm
The stop-and-go algorithm proposed by Picchi and Prati [11] can be seen as a variant of the Sato algorithm that achieves the same objectives of the BenvenisteGoursat algorithm with better convergence properties. The pseudo error for the stop-and-go algorithm is formulated as ( ek if sgn.ek / D sgn. S;k / (15.32) P;k D 0 otherwise where S;k is the Sato pseudo error given by (15.28), and ek is the error used in the decision-directed algorithm given by (15.29). The basic idea is that the algorithm converges if updating of the equalizer coefcients is turned off with sufciently high probability every time the sign of error (15.29) differs from the sign of error ei d;k D yk ak , that is sgn.ek / 6D sgn.ei d;k /. As ei d;k is not available in a self-training equalizer, with the stopand-go algorithm coefcient updating is turned off whenever the sign of error ek is different from the sign of Sato error S;k . Obviously, in this way we also get a non-zero probability that coefcient updating is inactive when the condition sgn.ek / D sgn.ei d;k / occurs, but this does not usually bias the convergence of the algorithm.
Remarks
At this point we can make the following observations. Self-training algorithms based on the minimization of a cost function that includes the term E[jyk j p ], p 2, can be explained referring to the algorithm (15.24), because the effect of raising to the p-th power the amplitude of the equalizer output sample is that of emphasizing the contribution of samples with large amplitude. Extension of the Sato cost function (15.25) to QAM systems, which we discuss in Section 15.5, is given by i h (15.33) J D E 1 jyk j2 S jyk j 2 where S D E[jak j2 ]=E[jak j]. In general this term guarantees that, at convergence, the statistical power of the equalizer output samples is equal to the statistical power of the input symbols.
1093
In the algorithm (15.24), the equalizer coefcients are updated only when we observe a peak value of the equalizer output signal. As the peak value decreases with the progress of the equalization process, updating of the equalizer coefcients ideally depends on a threshold that varies depending on the level of distortion in the overall system impulse response.
15.3
The algorithm (15.24) suggests that the equalizer coefcients are updated when the equalizer output sample reaches a threshold value, which in turn depends on the level of distortion in the overall system impulse response. In practice, we dene a threshold at instant k as Tk D max C C A;k , where the term C A;k 0 represents a suitable measure of distortion. Each time the absolute value of the equalizer output sample reaches or exceeds the threshold Tk , the coefcients are updated so that the peak value of the equalizer output is driven towards the constellation boundary max ; at convergence of the equalizer coefcients, C A;k vanishes. Figure 15.6 illustrates the evolution of the contour Tk for a two-dimensional constellation.
Im[ yk ] T0 Tk max
max
Re[ yk ]
1094
The updating of coefcients described above can be obtained by a stochastic gradient algorithm that is based on a cost function E[8.yk /] dened on the parameter space f i g. We assume that the overall system 9 initially corresponds to a point on a sphere of arbitrary radius r. With updating terms that on the average exhibit the same sign as the terms found in the general algorithm (15.24), the point on the sphere of radius r moves in such a way to reduce distortion. Moreover, if the radial component of the gradient, i.e. the component that is orthogonal to the surface of the sphere, is positive for r > 1, negative for r < 1 and vanishes on the sphere of radius r D 1, it is not necessary to scale the coefcients and the convergence will take place to the point of global minimum 9 D I . Clearly, the derivative function of 8.yk / with respect to yk that denes the pseudo error can be determined in various ways. A suitable denition is given by ( yk [max C C A;k ] sgn.yk / if jyk j max (15.34) 2.yk / D otherwise C A;k sgn.yk / For a self-training equalizer, the updating of coefcients as indicated by (15.8) is obtained by the algorithm ckC1 D ck 2.yk / xk (15.35)
In this algorithm, to avoid the computation of the threshold Tk at each iteration, the amplitude of the equalizer output signal is compared with max rather than with Tk ; note that the computation of 2.yk / depends on the event that yk falls inside or outside the constellation boundary. In the next section, for a two-dimensional constellation, we dene in general the constellation boundary as the contour line that connects the outer points of the constellation; for this reason we refer to this algorithm as the contour algorithm [12]. To derive the algorithms (15.34) and (15.35) from the algorithm (15.24) several approximations are introduced; consequently the convergence properties cannot be directly derived from those of the algorithm (15.24). In Appendix 15.A we show how C A should be dened to obtain the desired behavior of the algorithms, (15.34) and (15.35) in the case of systems with input symbols having a uniform continuous distribution. An advantage of the functional introduced in this section with respect to the Sato cost function is that the variance of the pseudo error vanishes at the points of minimum 9 D I ; this means that it is possible to obtain the convergence of the MSE to a steady state value that is close to the achievable minimum value. Furthermore, the radial component of the gradient of E[8.yk /] vanishes at every point on the unit sphere, whereas the radial component of the gradient in the Sato cost function vanishes on the unit sphere only at the points 9 D I . As the direction of steepest descent does not intersect the unit sphere, the contour algorithm avoids overshooting of the convergence trajectories observed using the Sato algorithm; in other words, the stochastic gradient yields a coefcient updating that is made in the correct direction more often than in the case of the Sato algorithm. Therefore, substantially better convergence properties are expected for the contour algorithm even in systems with a discrete probability distribution of input symbols. The complexity of the algorithms (15.34) and (15.35) can be deemed prohibitive for practical implementations, especially for self-training equalization in high-speed communication systems, as the parameter C A;k must be estimated at each iteration. In the next
1095
section, we discuss a simplied algorithm that allows implementation with low complexity; we will see later how the simplied formulation of the contour algorithm can be extended to self-training equalization of partial response and QAM systems.
h i ak
C wk
(15.36)
where fwk g denotes additive white Gaussian noise. The equalizer output is given by yk D T ck xk . To obtain an algorithm that does not require the knowledge of the parameter C A , the denition (15.34) suggests introducing the pseudo error ( yk max sgn.yk / if jyk j max (15.37) C A;k D k sgn.yk / otherwise where k is a non-negative parameter that is updated at every iteration as follows: 8 M 1 > > k 1 if jyk j max < M kC1 D (15.38) > > : k C 1 1 otherwise M and 1 is a positive constant. The initial value 0 is not a critical system parameter and can be, for example, chosen equal to zero; the coefcient updating algorithm is thus given by ckC1 D ck C A;k xk (15.39)
In comparison to (15.34), now k does not provide a measure of distortion as C A . The denition (15.37) is justied by the fact that the term yk [max C C A ] sgn.yk / in (15.34) can be approximated as yk max sgn.yk /, because if the event jyk j max occurs the pseudo error yk max sgn.yk / can be used for coefcient updating. Therefore, k should increase in the presence of distortion only in the case the event jyk j < max occurs more frequently than expected. This behavior of the parameter k is obtained by applying (15.38). Moreover, (15.38) guarantees that k assumes values that approach zero at the convergence of the equalization process; in fact, in this case the probabilities of the events fjyk j < max g and fjyk j max g assume approximately the values .M 1/=M and 1=M, respectively, that correspond to the probabilities of such events for a noisy PAM signal correctly equalized. Figure 15.7 shows the pseudo error C A;k as a function of the value of the equalizer output sample yk . The contour algorithm has been described for the case of uniformly distributed input symbols; however, this assumption is not necessary. In general, if fyk g represents an equalized signal, the terms .M 1/=M and 1=M in (15.38) are, respectively, substituted by p0 D P[jak j < max ] C
1 2
(15.40)
1096
CA,k
+k max k
+ max yk
Figure 15.7. Characteristic of the pseudo error CA,k as a function of the equalizer output.
and p1 D
1 2
(15.41)
We note that the considered receiver makes use of signal samples at the symbol rate; the algorithm can also be applied for the initial convergence of a fractionally spaced equalizer (see Section 8.4) in case a sampling rate higher than the symbol rate is adopted.
15.4
Self-training adaptive equalization has been mainly studied for full response systems; however, self-training equalization methods for partial response systems have been proposed for linear and non-linear equalizers [7, 13, 14]. In general, a self-training equalizer is more difcult to implement for partial response systems (see Appendix 7.A), especially if the input symbol alphabet has more than two elements. Moreover, as self-training is a slow process, the accuracy achieved in the recovery of the timing of the received signal before equalization plays an important role. In fact, if timing recovery is not accomplished, because of the difference between the sampling rate and the modulation rate we have that the sampling phase varies with respect to the timing phase of the remote transmitter clock; in this case we speak of phase drift of the received signal; self-training algorithms fail if the phase drift of the received signal is not sufciently small. In this section, we discuss the extension to partial response systems of the algorithms for PAM systems presented in the previous section.
1097
is given by .D/ D .1 D 2 /. The objective of an adaptive equalizer for a PR-IV system consists in obtaining an equalized signal of the form yk D .ak ak
2 / C w y;k
D u k C w y;k
(15.42)
where w y;k is a disturbance due to noise and residual distortion. We consider the case of quaternary modulation. Then the input symbols ak are from the set f 3; 1; C1, C3g, and the output signal u k D ak ak 1 , for an ideal channel in the absence of noise, can assume one of the seven values f 6; 4; 2; 0; C2; C4; C6g.6 As illustrated in Figure 15.8, to obtain a pseudo error to be employed in the equalizer coefcient updating algorithm, the equalizer output signal fyk g is transformed into a fullresponse signal v S;k by the linear transformation v S;k D yk C S v S;k
2
(15.43)
where S is a constant that satises the condition 0 < S < 1. Then the signal v S;k is quantized by a quantizer with two levels corresponding to S , where S is given by (15.26). The obtained signal is again transformed into a partial response signal that is subtracted from the equalizer output to generate the pseudo error S;k D yk
S [sgn.v S;k /
sgn.v S;k
2 /]
(15.44)
Figure 15.8. Block diagram of a self-training equalizer for a PR-IV system using the Sato algorithm.
In general, for an ideal PR-IV channel in the absence of noise, if the alphabet of the input symbols is A D f1; 3; : : : ; .M 1/g, the output symbols assume one of the .2M 1/ values f0; 2; : : : ; 2.M 1/g.
1098
Then the Sato algorithm for partial response systems is expressed as ckC1 D ck S;k xk (15.45)
C .1
C A / ak O
(15.46)
The equation (15.46) and the choice of the parameter C A are justied in the following way. If C A D 0 is selected, we obtain an equalization system with decision feedback, that presents the possibility of signicant error propagation. The effect of the choice C A D 1 is easily seen using the D transform. From (15.42) and assuming correct decisions, (15.46) can be expressed as v.D/ D a.D/ C w y .D/ 1 C A D 2 (15.47)
Therefore with C A D 1 we get the linear inversion of the PR-IV channel with innite noise enhancement at frequencies f D 0 and 1=.2T / Hz. The value of C A is chosen in the interval 0 < C A < 1, to obtain the best trade-off between linear feedback and decision feedback. We now apply the contour algorithm (15.37)(15.39) using the signal vk rather than the equalizer output signal yk . The pseudo error is dened as ( vk max sgn.vk / if jvk j max v (15.48) C A;k D v otherwise k sgn.vk /
v where k is a non-negative parameter that is updated at each iteration as 8 > v M 1 > 1 if jvk j max < k M v (15.49) kC1 D > v > C 1 1 : k otherwise M The stochastic gradient must be derived taking into consideration the channel equalization performed with linear feedback and decision feedback. We dene the error on the M-ary symbol ak after channel equalization as v ek D vk
ak
(15.50)
1099
Assuming correct decisions, by (15.46) it is possible to express the equalizer output as yk D .ak ak
v 2 / C ek v C A ek v ek 2 v C A ek 2
(15.51)
The equation (15.51) shows that an estimate of the term must be included as error signal in the expression of the stochastic gradient. After initial convergence of the O equalizer coefcients, the estimate ek D vk ak is reliable. Therefore decision-directed Ov coefcient updating can be performed according to the algorithm ckC1 D ck dd .ek Ov C A ek Ov
2 / xk
(15.52)
The contour algorithm for coefcient updating during self-training is obtained by substitutv v ing the decision-directed error signal ek with the pseudo error C A;k (15.48), ckC1 D ck
v .C A;k v C A k 2 / xk
(15.53)
During self-training, satisfactory convergence behavior is usually obtained for C A ' 1=2. Figure 15.9 shows the block diagram of an equalizer for a PR-IV system with a quaternary alphabet (QPR-IV), with the generation of the error signals to be used in decisiondirected and self-training mode.
Figure 15.9. Block diagram of a self-training equalizer with the contour algorithm for a QPR-IV system.
1100
In the described scheme, the samples of the received signal can be initially ltered by a lter with transfer function 1=.1 a D 2 /, 0 < a < 1, to reduce the correlation among samples. The obtained signal is then input to the equalizer delay line.
15.5
We now describe various self-training algorithms for passband transmission systems that employ a two-dimensional constellation.
(15.54)
and introduce the Sato cost function for QAM systems, J D E[8.yk;I / C 8.yk;Q /] where 8.v/ D
1 2
(15.56)
v2
jvj
(15.57)
Figure 15.10. Block diagram of a self-training equalizer with the Sato algorithm for a QAM system.
1101
and
S
2 E[ak;I ]
E[jak;I j]
2 E[ak;Q ]
E[jak;Q j]
(15.58)
(15.59)
sgn.v/
(15.60)
The partial derivative with respect to the carrier phase estimate is given by (see also (8.385)) @ J D E[Im.yk .2.yk;I / C j 2.yk;Q // /] @' O Dening the Sato pseudo error for QAM systems as S;k D yk and observing that
Im.yk S;k / D S
(15.61)
sgn.yk /
(15.62)
Im.
S yk
sgn.yk //
(15.63)
the equalizer coefcient updating and carrier phase estimate are given by ckC1 D ck
O S;k e j 'k x k S yk sgn.yk /]
(15.64) (15.65)
where ' is the adaptation gain of the carrier phase tracking loop. The equations (15.64) and (15.65) are analogous to (8.382) and (8.387) for the decision-directed case. The same observations made for self-training equalization of PAM systems using the Sato algorithm hold for QAM systems. In particular, assuming that the algorithm converges to a point of global minimum of the cost function, we recall that the variance of the pseudo error assumes high values in the neighborhood of the point of convergence. Therefore in the steady state it is necessary to adopt a decision directed algorithm. To obtain smooth transitions between the self-training mode and the decision-directed mode without the need of a further control of the distortion level, in QAM systems we can use extensions of the BenvenisteGoursat and stop-and-go algorithms considered for self-training of PAM systems.
15.5.1
In the constant modulus algorithm (CMA) for QAM systems proposed by Godard [15], self-training equalization is based on the cost function J D E[.j yk j p Q R p /2 ] D E[.jyk j p R p /2 ] (15.66)
1102
where p is a parameter that usually assumes the value p D 1 or p D 2. We note that, as J depends on the absolute value of the equalizer output raised to the p-th power, the CMA does not require the knowledge of the carrier phase estimate. Figure 15.11 shows the block diagram of a receiver using the CMA. The gradient of (15.66) is given by Q rc J D 2 p E[.j yk j p R p /j yk j p Q
2
yk x ] Q k
(15.67)
The constant R p is chosen so that the gradient is equal to zero for a perfectly equalized system; therefore we have Rp D E[jak j2 p ] E[jak j p ] (15.68)
For example, for a 64-QAM constellation, we obtain R1 D 6:9 and R2 D 58. The 64-QAM constellation and the circle of radius R1 D 6:9 are illustrated in Figure 15.12. By using (15.67), we obtain the equalizer coefcient updating law ckC1 D ck For p D 1, (15.69) becomes ckC1 D ck .j yk j Q R1 / yk Q x j yk j k Q (15.70) .j yk j p Q R p /j yk j p Q
2
yk x Q k
(15.69)
We note that the Sato algorithm, introduced in Section 15.2, can then be viewed as a particular case of the CMA.
1103
Im[ ak ]
R1 =6.9 R1 Re[ ak ]
Figure 15.12. The 64-QAM constellation and the circle of radius R1 D 6:9.
is still valid, and the complex-valued baseband equivalent channel output is given by xk D
C1 X iD 1
h i ak
C wk
(15.71)
The equalizer output is expressed as yk D cT xk . To generalize the notion of pseudo error of the contour algorithm introduced in Section 15.3 for PAM systems, we dene a contour line C that connects the outer points of the constellation. For simplicity, we assume a square constellation with L L points, as illustrated in Figure 15.13 for the case L D 8. Let S be the region of the complex plane enclosed by the contour line C and let C 2 S = C = by denition. We denote by yk the closest point to yk on C every time that yk 2 S, that is every time the point yk is found outside the region enclosed by C. The pseudo error (15.37) is now extended as follows: 8 C = se yk 2 S > yk yk < ) C A;k D (15.72) if jyk;I j jyk;Q j k sgn.yk;I / se yk 2 S > : j k sgn.yk;Q / if jyk;I j < jyk;Q j Also in this case k is a non-negative parameter, updated at each iteration as ( if yk 2 S = k pS 1 kC1 D k C .1 pS / 1 if yk 2 S
(15.73)
where, by analogy with (15.40), the probability pS ' P[yk 2 S] is computed assuming that yk is an equalized signal in the presence of additive noise.
1104
Im[ yk ]
~C yk
~ yk
Re[ yk ]
S C
Figure 15.13. Illustration of the contour line and surface S for a 64-QAM constellation.
Let max be the maximum absolute value of the real and imaginary parts of the square L L symbol constellation. If jyk;I j max or jyk;Q j max , but not both, the projection of the sample yk on the contour line C yields a non-zero pseudo error along one dimension C and a zero error in the other dimension. If both jyk;I j and jyk;Q j are larger than max , yk is chosen as the corner point of the constellation closest to yk ; in this case we obtain a non-zero pseudo error in both dimensions. Thus the equalizer coefcients are updated according to the algorithm ckC1 D ck C A;k x k (15.74)
Clearly, the contour algorithm can also be applied to systems that use non-square constellations. In any case, the robust algorithm for carrier phase tracking that is described in the next section requires that the shape of the constellation is non-circular.
(15.75)
As for equalizer coefcient updating, reliable information for updating the carrier phase estimate 'k is only available if yk falls outside of the region S. As illustrated in Figure 15.14, O
1105
(max ,+max )
Im[ yk ]
(+max ,+max )
yC k
k
yk
Figure 15.14. Illustration of the rotation of the symbol constellation in the presence of a phase error, and denition of 1'k .
the phase estimation error can then be computed as (see also (15.65))
C 1'k ' Im.yk yk / D
Im[yk .yk
C yk /]
(15.76)
If yk falls within S the phase error is set to zero, that is 1'k D 0. From Figure 15.14 we note that the phase error 1'k is invariant with respect to a rotation of yk equal to an integer multiple of =2. Then to determine 1'k we can rst rotate yk ,
0 yk D yk e j`=2
(15.77)
0 0 where ` is chosen such that Re[yn ] > jIm[yn ]j (shaded region in Figure 15.14). Furthermore, we observe that the information on the phase error obtained by samples of the sequence fyk g that fall in the corner regions, where jyk;I j > max and jyk;Q j > max , is not important. Thus we calculate a phase error only if yk is outside of S, but not in the corner regions, 0 0 0 0 that is if yk 2 D, with D D fyk : Re[yk ] > max ; jIm[yk ]j < max g. Then (15.76) becomes ( 0 0 0 Im[yk ][Re[yk ] max ] if yk 2 D (15.78) 1'k D 0 otherwise
In the presence of a frequency offset equal to =.2 /, the probability distribution of the samples fyk g rotates at a rate of =.2 / revolutions per second. For large values of =.2 /, the phase error 1'k does not provide sufcient information for the carrier phase tracking system to achieve a lock condition; therefore the update of 'k must be made by a secondO order PLL, where in the update of the second-order term a factor that is related to the value of the frequency offset must be included (see Section 14.7).
1106
0 The needed information is obtained observing the statistical behavior of the term Im[yk ], 0 2 D. At instants in which the sampling distribution of y is conditioned by the event yk k 0 aligned with S, the distribution of Im[yk ] is uniform in the range [ max ; max ]. Between 0 these instants, the distribution of Im[yk ] exhibits a time varying behavior with a downward or upward trend depending on the sign of the frequency offset, with a minimum variance when the corners of the rotating probability distribution of yk , which we recall rotates at a rate of =.2 / revolutions per second, cross the coordinate axes. Dening ( v if jvj < max (15.79) Q.v/ D 0 otherwise
from the observation of Figure 15.14 a simple method to extract information on consists in evaluating
0 0 1 Im[yk ] D QfIm[yk ] 0 Im[ym ]g 0 yk 2 D
=.2 / (15.80)
0 0 where m < k denotes the last time index for which yk 2 D. In the mean, 1 Im[yk ] exhibits the sign of the frequency offset. The equations for the updating of the parameters of a second-order phase-locked loop for the carrier phase recovery then become ( 'kC1 D 'k C ' 1'k C 1'c;k O O O 0 if yk 2 D 0 1'c;kC1 D 1'c;k C f 1 1'k C f 2 1 Im[yk ] O O (15.81) ( 'kC1 D 'k O O otherwise 1'c;kC1 D 1'c;k O O
where ' , f 1 and f 2 are suitable adaptation gains; typically, ' is in the range 10 4 10 3 , f 1 D .1=4/2 , and f 2 ' f 1 . ' 0 The rotation of yk given by (15.77) to obtain yk also has the advantage of simplifying the error computation for self-training equalizer coefcient adaptation with the contour algorithm. With no signicant effect on performance, we can introduce a simplication similar to that adopted to update the carrier phase, and let the pseudo error equal zero if yk is found 0 in the corner regions, that is C A;k D 0 if Im[yk ] > max . By using (15.72) and (15.73) to compute the pseudo error, the coefcient updating equation (15.74) becomes (see (8.382)) ckC1 D ck
O C A;k e j 'k x k
(15.82)
15.6
Examples of applications
In this section, we give examples of applications that illustrate the convergence behavior and steady state performance of self-training equalizers, with particular regard to the contour algorithm. We initially consider self-training equalization for PAM transmission systems over unshielded twisted-pair UTP cables with frequency response given by (4.169).
1107
Example 15.6.1 As a rst example, we consider a 16-PAM system (M D 16) with a uniform probability distribution of the input symbols and symbol rate equal to 25 MBaud; the transmit and receive lters are designed to yield an overall raised cosine channel characteristic for a cable length of 50 m. In the simulations, the cable length is chosen equal to 100 m, and the received signal is disturbed by additive white Gaussian noise. The signal-to-noise ratio at the receiver input is equal to 0 D 36 dB. Self-training equalization is achieved by a fractionally spaced equalizer having N D 32 coefcients, and input signal sampled with sampling period equal to T =2. Figure 15.15 shows the convergence of the contour algorithms (15.37) and
Figure 15.15. Illustration of the convergence of the contour algorithm for a 16-PAM system: (a) behavior of the parameter n , (b) MSE convergence, (c) relative frequency of equalizer output samples at the beginning and the end of the convergence process.
1108
(15.38) for 0 D 0 and c0 chosen equal to the zero vector. The results are obtained for a cable with attenuation (4.148) equal to . f / j f D1 D 3:85 10 6 [m 1 Hz 1=2 ], parameters of the self-training equalizer given by D 10 5 and 1 D 2:5 10 4 , and ideal timing recovery. Example 15.6.2 We consider self-training equalization for a baseband quaternary partial response class IV system (M D 4) for transmission at 125 Mbit/s over UTP cables; a VLSI transceiver implementation for this system will be described in Chapter 19. We compare the performance of the contour algorithm, described in Section 15.3, with the Sato algorithm for partial response systems. Various realizations of the MSE convergence of a self-training equalizer with N D 16 coefcients are shown in Figures 15.16 and 15.17 for the Sato algorithm and the contour algorithm, respectively. The curves are parameterized by t D 1T =T , where T D 16 ns, and 1T denotes the difference between the sampling phase of the channel output signal and the optimum sampling phase that yields the minimum MSE; we note that the contour algorithm has a faster convergence with respect to the Sato algorithm and yields signicantly lower values of MSE in the steady state. The Sato algorithm can be applied only if timing recovery is achieved prior to equalization; note that the convergence characteristics of the contour algorithm makes self-training equalization possible even in the presence of considerable distortion and phase drift of the received signal.
Figure 15.16. MSE convergence with the Sato algorithm for a QPR-IV system [13].
1109
Figure 15.17. MSE convergence with the contour algorithm for a QPR-IV system [13].
0.2
0.4 0 1 2 3 t/T 4 5 6
Figure 15.18. Overall baseband equivalent channel impulse response for simulations of a 256-QAM system. [From [12], c 1998 Springer-Verlag London, Ltd.]
1110
Example 15.6.3 We now examine self-training equalization for a 256-QAM transmission system having a square constellation with L D 16 (M D 256), and symbol rate equal to 6 MBaud. Along each dimension, symbols 3, 1 have probability 2=20, and symbols 15, 13, 11, 9, 7, and 5 have probability 1=20. The overall baseband equivalent channel impulse response is illustrated in Figure 15.18. The received signal is disturbed by additive white Gaussian noise. The signal-to-noise ratio at the receiver input is equal to 0 D 39 dB. Signal equalization is obtained by a fractionally spaced equalizer having N D 32 coefcients, and input signal sampled with sampling period equal to T =2. Figure 15.19 shows the convergence of the contour algorithm and the behavior of the parameter k for pS D 361=400, various initial values of 0 , and c0 given by a vector with all elements equal to zero except for one element. Results are obtained for D 10 4 , 1 D 10 4 , and ideal timing and carrier phase recovery. Example 15.6.4 With reference to the previous example, we examine the behavior of the carrier phase recovery algorithm, assuming ideal timing recovery. Figure 15.20 illustrates the behavior of the MSE and of the second-order term 1'c;k for an initial frequency offset of C2:5 kHz, O ' D 4 10 4 , f 1 D 8 10 8 , and f 2 D 2 10 8 .
Figure 15.19. Convergence behavior of MSE and parameter k using the contour algorithm for a 256-QAM system with non-uniform distribution of input symbols. [From [12], c 1998 Springer-Verlag London, Ltd.]
15. Bibliography
1111
Figure 15.20. Illustration of the convergence behavior of MSE and second-order term 1 ' c,k using the contour algorithm in the presence of an initial frequency offset equal to 500 ppm for a 256-QAM system with non-uniform distribution of input symbols. [From [12], c 1998 Springer-Verlag London, Ltd.]
Bibliography
[1] H. Ichikawa, J. Sango, and T. Murase, 256 QAM multicarrier 400 Mb/s microwave radio system eld tests, in Proc. 1987 IEEE Int. Conference on Communications, pp. 18031808, 1987. [2] F. J. Ross and D. P. Taylor, An enhancement to blind equalization algorithms, IEEE Trans. on Communications, vol. 39, pp. 636639, May 1991. [3] J. G. Proakis and C. L. Nikias, Blind equalization, in Proc. SPIE Adaptive Signal Processing, vol. 1565, pp. 7687, July 2224 1991. [4] S. Bellini, Blind equalization and deconvolution, in Proc. SPIE Adaptive Signal Processing, vol. 1565, pp. 88101, July 2224 1991. [5] N. Benvenuto and T. W. Goeddel, Classication of voiceband data signals using the constellation magnitude, IEEE Trans. on Communications, vol. 43, pp. 27592770, Nov. 1995.
1112
[6] R. Liu and L. Tong, eds, Special issue on blind systems identication and estimation, IEEE Proceedings, vol. 86, Oct. 1998. [7] Y. Sato, A method of self-recovering equalization for multilevel amplitudemodulation systems, IEEE Trans. on Communications, vol. 23, pp. 679682, June 1975. [8] L. Tong, G. Xu, B. Hassibi, and T. Kailath, Blind channel identication based on second-order statistics: a frequency-domain approach, IEEE Trans. on Information Theory, vol. 41, pp. 329334, Jan. 1995. [9] A. Benveniste, M. Goursat, and G. Ruget, Robust identication of a nonminimum phase system: blind adjustment of a linear equalizer in data communications, IEEE Trans. on Automatic Control, vol. 25, pp. 385399, June 1980. [10] A. Benveniste and M. Goursat, Blind equalizers, IEEE Trans. on Communications, vol. 32, pp. 871883, Aug. 1984. [11] G. Picchi and G. Prati, Blind equalization and carrier recovery using a Stop-and-Go decision directed algorithm, IEEE Trans. on Communications, vol. 35, pp. 877887, Sept. 1987. [12] G. Cherubini, S. Olcer, and G. Ungerboeck, The contour algorithm for self-training equalization, in Broadband Wireless Communications, 9th Tyrrhenian Int. Workshop on Digital Communications (M. Luise and S. Pupolin, eds), Lerici, Italy, pp. 5869, Sept. 710 1997. Berlin: Springer-Verlag, 1998. [13] G. Cherubini, S. Olcer, and G. Ungerboeck, Self-training adaptive equalization for multilevel partial-response transmission systems, in Proc. 1995 IEEE Int. Symposium on Information Theory, Whistler, Canada, p. 401, Sept. 1722 1995. [14] G. Cherubini, Nonlinear self-training adaptive equalization for partial-response systems, IEEE Trans. on Communications, vol. 42, pp. 367376, February/March/April 1994. [15] D. N. Godard, Self recovering equalization and carrier tracking in two-dimensional data communication systems, IEEE Trans. on Communications, vol. 28, pp. 1867 1875, Nov. 1980.
1113
Appendix 15.A
PC1 Given y D y0 D i D 1 ci x i , we show that the only minima of the cost function J D E[8.y/] correspond to the equalizer settings C D H 1 , for which we obtain 9 D I , except for a possible delay. If the systems H and C have nite energy, then also 9 has nite energy and P E[8.y/] may be regarded as a functional J .9/ D E[8.y/], C1 where y is expressed as y D i D 1 i a i ; thus we must prove that the only minima of V.9/ are found at points 9 D I . We consider input symbols with a probability density function pak ./ uniform in the interval [ max ; max ]. N N For the analysis, we express the system 9 as 9 D r 9, where r 0, and 9 $ f N i g, P with i N i2 D 1, denotes the normalized overall system. We consider the cost function J PC1 as a functional V.9/ D E[8.r y /], where y D y0 D i D 1 N i a i denotes the output of N N N N the system 9, and 8 has derivative 2 given by (15.34). Let ( x max sgn.x/ if jxj max Q 2.x/ D (15.83) 0 otherwise and p yk .x/ denote the probability density function of y . N N We express the parameter C A as
CA
N 9
Z
N 9
(15.84)
N N Examine the function V.9/ on the unit sphere S. To claim that the only minima of V.9/ are found at points 9 D I , we apply Theorem 3.5 of [9]. Consider a pair of indices .i; j/, P N2 i 6D j, and a xed system with coefcients f N ` g`6Di; j , such that R 2 D 1 `6Di; j ` > 0. N For ' 2 [0; 2 /, let 9' 2 S be the system with coefcients f N ` g`6Di; j , N i D R cos ', and N j D R sin '; moreover, let .@=@'/V.9' / be the derivative of V.9' / with respect to ' at N N N N point 9 D 9' . As pak ./ is sub-Gaussian, it can be shown that @ N V.9' / D 0 @' and @ N V.9' / > 0 @' for 0 < ' < 4 (15.86) for ' D k 4 k2Z (15.85)
1114
N From the above equations we have that the stationary points of V.9' / correspond to systems characterized by the property that all non-zero coefcients have the same absolute value. Furthermore, using symmetries of the problem, we nd the only minima are at I , except N for a possible delay, and the other stationary points of V.9' / are saddle points. The study of the functional V is then extended to the entire parameter space. As the results obtained for the restriction of V to S are also valid on a sphere of arbitrary radius r, we need to study only the radial derivatives of V. For this reason, we consider the function Q N V.r/ D V.r 9/, whose rst and second derivatives are Z Z Q 0 .r/ D b 2.r b/ p yk .b/ db . N C r / jbj p yk .b/ db Q V (15.87) N N 9 and Q V 00 .r/ D Z Q b 20 .r b/ p yk .b/ db N
2 0 r
jbj p yk .b/ db N
(15.88)
Q where 20 and r0 denote derivatives. Q Recalling the expressions of 9 and r given by (15.84), we obtain V 0 .0/ < 0 and N Q 00 .r/ > 0. Therefore there exists a radius r0 such that the radial component of the gradient V N is negative for r < r0 and positive for r > r0 . For a xed point 9 2 S, r0 is given by the solution of the equation Z Z Q (15.89) b 2.r b/ p yk .b/ db . 9 C r / jbj p yk .b/ db D 0 N N N N Substituting the expressions of 9 and r in (15.89), we obtain r0 D 1, 89 2 S. Therefore N the only minima of V are at I . Furthermore, as the radial component of the gradient vanishes on S, the steepest descent lines of V do not cross the unit sphere. Using the same argument given in [9], we conclude that the points I are the only stable attractors of the steepest descent lines of the function V, and that the unique stable attractors of the steepest descent lines of J are H 1 . Note that the parameter 9 is related to the distortion of the distribution of the input N sequence ltered by a normalized system; this parameter varies along the trajectories of the stochastic gradient algorithm and vanishes as a point of minimum is reached. Moreover, note that the parameter r indicates the deviation of the overall system gain from the desired unit value; if the gain is too small, the gradient is augmented with an additional driving term.
Chapter 16
The algorithms and structures discussed in this chapter can be applied to both wired and wireless systems, even though transmission systems over twisted-pair cables will be considered to describe examples of applications. Full-duplex data transmission over a single twisted-pair cable permits the simultaneous ow of information in two directions using the same frequency band. Examples of applications of this technique are found in digital communications systems that operate over the telephone network. In a digital subscriber loop, at each end of the full-duplex link, a circuit called hybrid separates the two directions of transmission. To avoid signal reections at the near and far-end hybrid, a precise knowledge of the line impedance would be required. As the line impedance depends on line parameters that, in general, are not exactly known, an attenuated and distorted replica of the transmit signal leaks to the receiver input as an echo signal. Data-driven adaptive echo cancellation mitigates the effects of impedance mismatch. A similar problem is caused by cross-talk in transmission systems over voice-grade unshielded twisted-pair cables for local-area network applications, where multipair cables are used to physically separate the two directions of transmission. Cross-talk is a statistical phenomenon due to randomly varying differential capacitive and inductive coupling between adjacent two-wire transmission lines (see Section 4.4.2). At the rates of several megabit-persecond that are usually considered for local-area network applications, nearend cross-talk (NEXT) represents the dominant disturbance; hence adaptive NEXT cancellation must be performed to ensure reliable communications. A different problem shows up in frequencydivision duplexing transmission, where different frequency bands are used in the two directions of transmission. In such a case farend cross-talk (FEXT) interference is dominant. This situation occurs, for instance, in very highspeed digital subscriber line (VDSL) systems, where multiple users are connected to a central station via unshielded twisted-pairs located in the same cable binder. In voiceband data modems the model for the echo channel is considerably different from the echo channel model adopted for baseband transmission. The transmitted signal is a passband QAM signal, and the far-end echo may exhibit signicant carrier-phase jitter and carrier-frequency shift, which are caused by signal processing at intermediate points in the telephone network. Therefore a digital adaptive echo canceller for voiceband modems needs to embody algorithms that account for the presence of such additional impairments.
1116
In the rst three sections of this chapter,1 we describe the echo channel models and adaptive echo canceller structures for various digital communications systems, which are classied according to the employed modulation techniques. We also address the trade-offs between complexity, speed of adaptation, and accuracy of cancellation in adaptive echo cancellers. In the last section, we address the problem of FEXT interference cancellation for upstream transmission in a VDSL system. The system is modelled as a multiinput multioutput (MIMO) system, to which multiuser detection techniques (see Section 10.4) can be applied.
16.1
The model of a full-duplex baseband PAM data transmission system employing adaptive echo cancellation is shown in Figure 16.1. To describe system operations, we consider one end of the full-duplex link. The conguration of an echo canceller for a PAM transmission system (see Section 6.13) is shown in Figure 16.2. The transmitted data consist of a sequence fak g of i.i.d. real-valued symbols from the M-ary alphabet A D f1, 3; : : : ; .M 1/g. The sequence fak g is converted into an analog signal by a digital-to-analog (D/A) converter (see Chapter 7). The conversion to a staircase signal by a zero-order hold D/A converter is described by the frequency response HD=A . f / D T sin. f T /=. f T /, where T denotes the modulation interval. The D/A converter output is ltered by the analog transmit lter and is input to the channel through the hybrid. The signal x.t/ at the output of the low-pass analog receive lter has three components, namely, the signal from the far-end transmitter r.t/, the echo u.t/, and additive Gaussian noise w R .t/. The signal x.t/ is given by x.t/ D r.t/ C u.t/ C w R .t/ D
C1 X kD 1 R ak h.t
kT / C
C1 X kD 1
ak h E .t
kT / C w R .t/ (16.1)
R where fak g is the sequence of symbols from the remote transmitter, and h.t/ and h E .t/ D fh D=A g E g.t/ are the impulse responses of the overall channel and of the echo channel,
The material presented in Sections 16.116.3 is reproduced with permission from G. Cherubini, Echo Cancellation, in The Mobile Communications Handbook (J. D. Gibson, ed.), Ch. 7, pp. 7.17.15, Boca Raton, FL: CRC Press, 1999, 2nd ed. c 1999 CRC Press, Boca Raton, FL.
1117
respectively. In the expression of h E .t/, the function h D=A .t/ is the inverse Fourier transform of HD=A . f /: The signal obtained after echo cancellation is processed by a detector that outputs the sequence of detected symbols fak g. OR
1118
rate R b /2 H
Tx rate R b
form of interpolation (see Chapter 14) would be required that can signicantly increase the transceiver complexity.
i D 0; : : : ; F0 C wk F0 Ci
C1 X nD 1
h E;n F0 Ci ak
(16.2)
where fh k F0 Ci ; i D 0; : : : ; F0 1g and fh E;k F0 Ci ; i D 0; : : : ; F0 1g are the discretetime impulse responses of the overall channel and the echo channel, respectively, and fwk F0 Ci ; i D 0; : : : ; F0 1g is a sequence of Gaussian noise samples with zero mean and 2 variance w . Equation (16.2) suggests that the sequence of samples fx k F0 Ci ; i D 0; : : : ; F0 1g be regarded as a set of F0 interlaced sequences, each with a sampling rate equal to the modulation rate. Similarly, the sequence of echo samples fu k F0 Ci ; i D 0; : : : ; F0 1g can be regarded as a set of F0 interlaced sequences that are output by F0 independent echo channels with discrete-time impulse responses fh E;k F0 Ci g; i D 0; : : : ; F0 1, and an identical sequence fak g of input symbols [1]. Hence, echo cancellation can be performed by F0 interlaced echo cancellers, as shown in Figure 16.4. As the performance of each
1119
{ak }
E.C. (0) ^ {u k F }
0
E.C. (1) ^ {u k F +1 } 0
E.C. ( F0 1) ^ {u k F + F } 0 0 t=l T F0
canceller is independent of the other F0 1 units, in the remaining part of this section we will consider the operations of a single echo canceller.
cn;k ak
(16.3)
The estimate of the echo is subtracted from the received signal. The result is dened as the cancellation error signal z k D xk u k D xk O
T c k ak
(16.4)
The echo attenuation that must be provided by the echo canceller to achieve proper system operation depends on the application. For example, for the integrated services digital network (ISDN) U-Interface transceiver, the echo attenuation must be larger than 55 dB [2]. It is then required that the echo signals outside of the time span of the echo canceller delay line be negligible, i.e., h E;n 0 for n < 0 and n > N 1: As a measure of system performance, we consider the mean-square error Jk at the output of the echo canceller at time t D kT , dened by
2 Jk D E[z k ]
(16.5)
1120
For a particular coefcient vector ck , substitution of (16.4) into (16.5) yields (see (2.17))
2 Jk D E[x k ] T T 2ck p C ck Rck
(16.6)
T where p D E[x k ak ] and R D E[ak ak ]. With the assumption of i.i.d. transmitted symbols, the correlation matrix R is diagonal. The elements on the diagonal are equal to the variance 2 of the transmitted symbols, a D .M 2 1/=3. From (2.40) the minimum mean-square error is given by 2 Jmin D E[x k ] T copt Rcopt
(16.7)
where the optimum coefcient vector is copt D R 1 p. We note that proper system operation is achieved only if the transmitted symbols are uncorrelated with the symbols from the remote transmitter. If this condition is satised, the optimum lter coefcients are given by the values of the discrete-time echo channel impulse response, i.e., cn;opt D h E;n , n D 0; : : : ; N 1.
Adaptive canceller
By the LMS algorithm, the coefcients of the echo canceller converge in the mean to copt . The LMS algorithm (see Section 3.1.2) for an N -tap adaptive linear transversal lter is formulated as follows: ckC1 D ck C z k ak (16.8)
where is the adaptation gain. The block diagram of an adaptive transversal lter echo canceller is shown in Figure 16.5. If we dene the vector ck D ck copt , the mean-square error can be expressed by (2.40) Jk D Jmin C
T ck R ck
(16.9)
T where the term ck R ck represents an excess mean-square error due to the misadjustment of the lter settings. Under the assumption that the vectors ck and ak are statistically independent, the dynamics of the mean-square error are given by (see (3.272)) 2 Jk D 0 [1 2 a .2 2 N a /]k C
2Jmin 2 2 N a
(16.10)
2 where 0 is determined by the initial conditions. The mean-square error converges to a 2 nite steady-state value J1 if the stability condition 0 < < 2=.N a / is satised. The optimum adaptation gain that yields fastest convergence at the beginning of the adaptation 2 process is opt D 1=.N a /. The corresponding time constant and asymptotic mean-square error are opt D N and J1 D 2Jmin , respectively. We note that a xed adaptation gain equal to opt could not be adopted in practice, as after echo cancellation the signal from the remote transmitter would be embedded in a residual echo having approximately the same power. If the time constant of the convergence
1121
ak
ak T
a k1 T
a kN+2
a kN+1
c0,k
c1,k
cN2,k
c N1,k
zk
+ ^ uk
+
xk
mode is not a critical system parameter, an adaptation gain smaller than opt will be adopted to achieve an asymptotic mean-square error close to Jmin . On the other hand, if fast convergence is required, a variable adaptation gain will be chosen. Several techniques have been proposed to increase the speed of convergence of the LMS algorithm. In particular, for echo cancellation in data transmission, the speed of adaptation is reduced by the presence of the signal from the remote transmitter in the cancellation error. To mitigate this problem, the data signal can be adaptively removed from the cancellation error by a decision-directed algorithm [3]. Modied versions of the LMS algorithm have been also proposed to reduce system complexity. For example, the sign algorithm suggests that only the sign of the error signal be used to compute an approximation of the gradient [4]. An alternative means to reduce the implementation complexity of an adaptive echo canceller consists in the choice of a lter structure with a lower computational complexity than the transversal lter.
1122
implementation, we consider echo cancellation by a distributed-arithmetic lter, where multiplications are replaced by table look-up and shift-and-add operations [5]. By segmenting the echo canceller into lter sections of shorter lengths, various trade-offs concerning the number of operations per modulation interval and the number of memory locations needed to store the look-up tables are possible. Adaptivity is achieved by updating the values stored in the look-up tables by the LMS algorithm. To describe the principles of operations of a distributed-arithmetic echo canceller, we assume that the number of elements in the alphabet of the input symbols is a power of two, M D 2W . Therefore, each symbol .0/ .W .i is represented by the vector [ak ; : : : ; ak 1/ ], where ak / 2 f0; 1g, i D 0; : : : ; W 1, are independent binary random variables, i.e., ak D
W 1 X wD0 .w/ .2ak
1/ 2w D
W 1 X wD0
.w/ bk 2w
(16.11)
.w/ .w/ 1/ 2 f 1; C1g. where bk D .2ak By substituting (16.11) into (16.3) and segmenting the delay line of the echo canceller into L sections with K D N =L delay elements each, we obtain " # L 1 X K 1 XW 1 X .w/ w 2 bk `K m c`K Cm;k uk D O (16.12) `D0 wD0 mD0
Note that the summation within parenthesis in (16.12) may assume at most 2 K distinct real .w/ values, one for each binary sequence fak `K m g, m D 0; : : : ; K 1. If we precompute these 2 K values and store them in a look-up table addressed by the binary sequence, we can substitute the real time summation by a simple reading from the table. Equation (16.12) suggests that the lter output can be computed using a set of L2 K values that are stored in L tables with 2 K memory locations each. The binary vectors .w/ .w/ a.w/ D [ak `K ; : : : ; ak `K K C1 ], w D 0; : : : ; W 1, ` D 0; : : : ; L 1, determine the k;` addresses of the memory locations where the values that are needed to compute the lter output are stored. The lter output is obtained by W L table look-up and shift-and-add operations. N k;` We observe that a.w/ and its binary complement a.w/ select two values that differ only k;` in their sign. This symmetry is exploited to halve the number of values to be stored. To determine the output of a distributed-arithmetic lter with reduced memory size, we reformulate (16.12) as " # K 1 L 1 X XW 1 X .w/ .w/ .w/ 2w bk `K c`K ;k C bk `K bk `K m c`K Cm;k (16.13) uk D O
`D0 wD0 mD1 .w/ Then the binary symbol bk `K determines whether a selected value is to be added or subtracted. Each table has now 2 K 1 memory locations, and the lter output is given by L 1 X XW 1 `D0 wD0 .w/ .w/ 2w bk `K dk .i k;` ; `/
uk D O
(16.14)
1123
m c`K Cm;k
1; ` D 0; : : : ; L 1, are the look-up indices computed as follows: 8 > K 1 .w/ > X .w/ > ak `K m 2m 1 if ak `K D 1 > < .w/ (16.15) i k;` D mD1 > K 1 .w/ > X .w/ m 1 > ak `K m 2 N if ak `K D 0 > :
mD1
We note that, as long as (16.12) and (16.13) hold for some coefcient vector [c0;k ; : : : ; c N 1;k ], a distributed-arithmetic lter emulates the operation of a linear transversal lter. For arbitrary values dk .n; `/, however, a non-linear ltering operation results. The expression of the LMS algorithm to update the values of a distributed-arithmetic echo canceller is derived as in (3.280). To simplify the notation we set u k .`/ D O hence (16.14) can be written as uk D O
L 1 X `D0 W 1 X wD0 .w/ .w/ 2w bk `K dk .i k;` ; `/
(16.16)
u k .`/ O
(16.17)
We also dene the vector of the values in the `-th look-up table as dk .`/ D [dk .0; `/; : : : ; dk .2 K
1
1; `/]T
(16.18)
indexed by the variable (16.15). The values dk .`/ are updated according to the LMS criterion, i.e., dkC1 .`/ D dk .`/ where
2 rdk .`/ z k D 2z k rdk .`/ z k D 1 2 2 rdk .`/ z k
(16.19)
2z k rdk .`/ u k D O
(16.20)
The last expression has been obtained using (16.17) and the fact that only u k .`/ depends O on dk .`/. Dening yk .`/ D [yk .0; `/; : : : ; yk .2 K (16.19) becomes dkC1 .`/ D dk .`/ C z k yk .`/ (16.21)
1
1124
For a given value of k and `, we assign the following values to the W addresses (16.15):
.w/ I .w/ D i k;`
w D 0; 1; : : : ; W
(16.22)
In conclusion, in (16.21) for every instant k and for each value of the index w D .w/ 0; 1; : : : ; W 1, the product 2w bk `K z k is added to the memory location indexed by I .w/ . The complexity of the implementation can be reduced by updating, at every iteration k, only the values corresponding to the addresses given by the most signicant bits of the symbols in the lter delay line. In this case (16.22) simplies into ( yk .n; `/ D 2W 0
1 b.W 1/ k `K
n D I .W n 6D
1/
I .W 1/
(16.23)
The block diagram of an adaptive distributed-arithmetic echo canceller with input symbols from a quaternary alphabet is shown in Figure 16.6. The analysis of the mean-square error convergence behavior and steady-state performance can be extended to adaptive distributed-arithmetic echo cancellers [6]. The dynamics of the mean-square error are in this case given by Jk D
2 0
2 a .2 2K 1
2 La /
k C
2Jmin 2 2 La
(16.24)
2 The stability condition for the echo canceller is 0 < < 2=.La /. For a given adaptation gain, echo canceller stability depends on the number of tables and on the variance of the transmitted symbols. Therefore, the time span of the echo canceller can be increased without affecting system stability, provided that the number L of tables is kept constant. In that case, however, mean-square error convergence will be slower. From (16.24), we nd that the optimum adaptation gain that permits the fastest mean-square error convergence 2 at the beginning of the adaptation process is opt D 1=.La /. The time constant of the K 1 . The smallest achievable time constant is therefore convergence mode is opt D L2 proportional to the total number of values. As mentioned above, the implementation of a distributed-arithmetic echo canceller can be simplied by updating at each iteration only the values that are addressed by the most signicant bits of the symbols stored in the delay line. The complexity required for adaptation can thus be reduced at the price of a slower rate of convergence.
16.2
Although most of the concepts presented in the preceding section can be readily extended to echo cancellation for communications systems employing QAM, the case of
1125
ak(0)
ak(1)
address computation
(1) i k,0 (0) ik,0
(0) a k(L1)K
a k(L1)K
(1)
address computation
(1) i k,L1 (0) i k,L1
zk
(1) d k+1 (i k,0
table
,0)
zk
table L1 +
0
(0) d k (i k,0 ,0)
1 0 1 0
+1 1 +1 1
bk(1)
1 0 1 0 2
+1 1 +1 1
b k(L1)K
(1)
bk(0)
zk
^ uk
full-duplex transmission over a voiceband data channel requires a specic discussion. We consider the system model shown in Figure 16.7. The transmitter generates a sequence fak g of i.i.d. complex-valued symbols from a two-dimensional constellation A, that are modulated by the carrier e j2 f 0 kT , where T and f 0 denote the modulation interval and the carrier frequency, respectively. The discrete-time signal at the output of the transmit passband phase splitter lter may be regarded as an analytic signal, which is generated at the rate of Q 0 =T samples/s, Q 0 > 1. The real part of the analytic signal is converted into an analog signal by a D/A converter and input to the channel. We note that by transmitting the real part of a complex-valued signal, positive and negativefrequency components become folded. The attenuation in the image band of the transmit lter thus determines the achievable echo suppression. In fact, the receiver cannot extract aliasing image-band components from desired passband frequency components, and the echo canceller is able to suppress only echo arising from transmitted passband components.
(0) b k(L1)K
xk
1126
The output of the echo channel is represented as the sum of two contributions. The near-end echo u N E .t/ arises from the impedance mismatch between the hybrid and the transmission line, as in the case of baseband transmission. The far-end echo u F E .t/ represents the contribution due to echoes that are generated at intermediate points in the telephone network. These echoes are characterized by additional impairments, such as jitter and frequency shift, which are accounted for by introducing a carrier-phase rotation equal to '.t/ in the model of the far-end echo. At the receiver, samples of the signal at the channel output are obtained synchronously with the transmitter timing, at the sampling rate of Q 0 =T samples/s. The discrete-time received signal is converted to a complex-valued baseband signal fx k F0 Ci g, i D 0; : : : ; F0 1, at the rate of F0 =T samples/s, 1 < F0 < Q 0 , through ltering by the receive phase splitter lter, decimation, and demodulation. From delayed transmit symbols, estimates of the near and far-end echo signals after demodulation, fu k FE Ci g, i D 0; : : : ; F0 1, and ON 0 F E g, i D 0; : : : ; F fu k F0 Ci O 0 1, respectively, are generated using F0 interlaced near and far-end echo cancellers. The cancellation error is given by z ` D x` .u ` E C u ` E / ON OF (16.25)
A different model is obtained if echo cancellation is accomplished before demodulation. In this case, two equivalent congurations for the echo canceller may be considered. In one conguration, the modulated symbols are input to the transversal lter, which approximates the passband echo response. Alternatively, the modulator can be placed after the transversal lter, which is then called a baseband transversal lter [7].
1127
In the considered implementation, the estimates of the echo signals after demodulation are given by u k FE Ci D ON 0
NX 1 NE nD0 NE cn F0 Ci;k ak n
i D 0; : : : ; F0 !
(16.26)
u k F0 Ci D OFE
NX 1 FE nD0
FE cn F0 Ci;k ak
n DE
O e j 'k F0 C1
i D 0; : : : ; F0
(16.27)
NE N FE F where [c0;k ; : : : ; c F0E N E 1;k ] and [c0;k ; : : : ; c F0E F E 1;k ] are the coefcients of the F0 inN N terlaced near and far-end echo cancellers, respectively, f'k F0 Ci g, i D 0; : : : ; F0 1, is the O sequence of far-end echo phase estimates, and D F E denotes the bulk delay accounting for the round-trip delay from the transmitter to the point of echo generation. To prevent overlap of the time span of the near-end echo canceller with the time span of the far-end echo canceller, the condition D F E > N N E must be satised. We also note that, because of the different nature of near and far-end echo generation, the time span of the far-end echo canceller needs to be larger than the time span of the near-end canceller, i.e., N F E > N N E . Adaptation of the lter coefcients in the near and far-end echo cancellers by the LMS algorithm leads to NE NE cn F0 Ci;kC1 D cn F0 Ci;k C z k F0 Ci .ak n/
n D 0; : : : ; N N E and
FE FE cn F0 Ci;kC1 D cn F0 Ci;k C z k F0 Ci .ak
i D 0; : : : ; F0
(16.28)
n DF E /
j 'k F0 Ci O
n D 0; : : : ; N F E
i D 0; : : : ; F0
(16.29)
respectively. The far-end echo phase estimate is computed by a second-order phase-lock loop algorithm (see Section 14.7), where the following gradient approach is adopted: ( O '`C1 D '` 1 ' r' jz ` j2 C 1'` .mod 2 / O O 2 (16.30) 1 1'`C1 D 1'` 2 f r' jz ` j2 O where ` D k F0 C i, i D 0; : : : ; F0 r' jz ` j2 D O 1, ' and f are parameters of the loop, and @jz ` j2 D @ '` O 2 Imfz ` .u ` E / g OF (16.31)
We note that the algorithm (16.30) requires F0 iterations per modulation interval, i.e., we cannot resort to interlacing to reduce the complexity of the computation of the far-end echo phase estimate.
1128
16.3
We discuss echo cancellation for OFDM with reference to a DMT system (see Chapter 9), as shown in Figure 16.8. Let fh i g, i D 0; : : : ; Nc 1, be the channel impulse response, sampled with period T =M, with length Nc M and fh E;i g, i D 0; : : : ; N 1, be the discrete-time echo impulse response whose length is N < M, where M denotes the number of subchannels of the DMT system. To simplify the notation, the length of the cyclic prex will be set to L D Nc 1. Recall that in a DMT transmitter the block of M samples at the IDFT output in the k-th modulation interval, [Ak [0]; : : : ; Ak [M 1]], is cyclically extended by copying the last L samples at the beginning of the block. After a P/S conversion, wherein the L samples of the cyclic extension are the rst to be output, the L C M samples of the block are transmitted into the channel. At the receiver the sequence is split into blocks of length L C M, [x k.MCL/ ; : : : ; x .kC1/.MCL/ 1 ]. These blocks are separated in such a way that the last M samples depend only on a single cyclically extended block, then the rst L samples are discarded.
Case N L C1. We initially assume that the length of the echo channel impulse response is N L C 1: Furthermore, we assume that the boundaries of the received blocks are placed such that the last M samples of the k-th received block are expressed by the vector (see (9.72)) xk D
R k
hC
h E C wk
(16.32)
1129
where h D [h 0 ; : : : ; h L ; 0; : : : ; 0]T is the vector of the overall channel impulse response extended with M L 1 zeros, h E D [h E;0 ; : : : ; h E;N 1 ; 0; : : : ; 0]T is the vector of the overall echo impulse response extended with M N zeros, wk is a vector of AWGN R samples, k is the circulant matrix with elements given by the signals from the remote transmitter 3 2 R R R Ak [0] Ak [M 1] : : : Ak [1] R R R 6 Ak [1] Ak [0] : : : Ak [2] 7 7 6 R (16.33) D6 : : : 7 k : : : 5 4 : : :
R Ak [M R 1] Ak [M R 2] : : : Ak [0]
and
is the circulant matrix with elements generated by the local transmitter 3 2 Ak [0] Ak [M 1] : : : Ak [1] 6 Ak [0] : : : Ak [2] 7 Ak [1] 7 6 : : : 7 k D6 : : : 5 4 : : : Ak [M 1] Ak [M 2] : : : Ak [0]
(16.34)
where H E denotes the DFT of the vector h E . In this case, the echo canceller provides an echo estimate that is given by O Uk D diag.ak / Ck (16.36)
where Ck denotes the DFT of the vector ck of the N coefcients of the echo canceller lter extended with M N zeros. In the time domain, (16.36) corresponds to the estimate O uk D
k ck
(16.37)
Case N > L C 1. In practice, however, we need to consider the case N > L C 1. The expression of the cancellation error is then given by xk D where
k;k 1 k;k 1 R k
hC
k;k 1 h E
C wk
(16.38)
2 6 6 6 4
Ak [M
1] Ak [M
2] : : :
1130
From (16.38) the expression of the cancellation error in the time domain is then given by z k D xk
k;k 1 ck
(16.40)
k;k 1
(16.41)
(16.42)
diag.ak / Ck
(16.43)
Equation (16.43) suggests a computationally efcient, two-part echo cancellation technique. First, in the time domain, a short convolution is performed and the result subtracted from the received signals to compensate for the insufcient length of the cyclic extension. Second, in the frequency domain, cancellation of the residual echo is performed over a set of M independent echo subchannels. Observing that (16.43) is equivalent to Z k D Xk Q k;k
1 Ck
(16.44)
1 where Q k;k 1 D FM k;k 1 FM ; the echo canceller adaptation by the LMS algorithm in the frequency domain takes the form H CkC1 D Ck C Q k;k 1 Zk
(16.45)
where is the adaptation gain. We note that, alternatively, echo canceller adaptation may also be performed by the simplied algorithm [8] CkC1 D Ck C diag.a / Zk k (16.46)
which entails a substantially lower computational complexity than the LMS algorithm, at the price of a slower rate of convergence. In DMT systems it is essential that the length of the channel impulse response be much less than the number of subchannels, so that the reduction in rate due to the cyclic extension may be considered negligible. Therefore, time-domain equalization is adopted in practice to shorten the length of the channel impulse response. From (16.43), however, we observe that transceiver complexity depends on the relative lengths of the echo and of the channel impulse responses. To reduce the length of the cyclic extension as well as the computational complexity of the echo canceller, various methods have been proposed to shorten both the channel and the echo impulse responses jointly [9].
1131
16.4
In this section, we address the problem of multiuser detection for upstream VDSL transmission (see Chapter 17), where FEXT signals at the input of a VDSL receiver are viewed as interferers that share the same channel as the remote user signal [10]. We assume knowledge of the FEXT responses at the central ofce and consider a decision-feedback equalizer (DFE) structure with cross-coupled linear feedforward (FF) equalizers and feedback (FB) lters for cross-talk suppression. DFE structures with crosscoupled lters have also been considered for interference suppression in wireless CDMA communications [11] and fast Ethernet transmission (see Appendix 19.A). Here we determine the optimum DFE coefcients in a minimum mean-square error (MMSE) sense assuming that each user adopts OFDM modulation for upstream transmission. A system with reduced complexity may be considered for practical applications, in which for each user and each subchannel only the most signicant interferers are suppressed. To obtain a receiver structure for multiuser detection that exhibits moderate complexity, we assume that each user adopts FMT modulation with M subchannels for upstream transmission (see Chapter 9). Hence the subchannel signals exhibit non-zero excess bandwidth as well as negligible spectral overlap. Assuming upstream transmission by U users, the system illustrated in Figure 16.9 is considered. In general, the sequences of subchannel signal samples at the multicarrier demodulator output are obtained at a sampling rate equal to a rational multiple F0 of the modulation rate 1=T . To simplify the analysis, here an integer F0 2 is assumed. We introduce the following denitions:
.u/ 1. fak [i]g, sequence of i.i.d. complex-valued symbols from a QAM constellation A.u/ [i] 2 transmitted by user u over subchannel i, with variance a .u/ [i ] ;
2. fak [i]g, sequence of detected symbols of user u at the output of the decision element O .u/ of subchannel i;
.u/ 3. h n [i], overall impulse response of subchannel i of user u;
4. h .u;v/ [i], overall FEXT response of subchannel i, from user v to user u; FEXT;n 5. G .u/ [i], gain that determines the power of the signal of user u transmitted over subchannel i; 6. fwn [i]g, sequence of additive Gaussian noise samples with correlation function Q .u/ rw.u/ [i ] .m/.
1132
Figure 16.9. Block diagram of transmission channel and DFE structure. [From [10], c 2001 IEEE.]
At the output of subchannel i of the user-u demodulator, the complex baseband signal is given by
.u/ xn [i] D G .u/ [i] 1 X kD 1 U X v6Du
vD1
G .v/ [i]
h .u;v/ FEXT;n
(16.47)
For user u, symbol detection at the output of subchannel i is achieved by a DFE structure such that the input to the decision element is obtained by combining the output signals of U linear lters and U feedback lters from all users, as illustrated in Figure 16.9. In practice, to reduce system complexity, for each user only a subset of all other user signals (interferers) is considered as an input to the DFE structure [10]. The selection of the subset signals is based on the power and the number of the interferers. This strategy, however, results in a loss of performance, as some strong interferers may not be considered.
1133
This effect is similar to the nearfar problem in CDMA systems. To alleviate this problem, it is necessary to introduce power control of the transmitted signals; a suitable method will be described in the next section. To determine the DFE lter coefcients, we assume M1 and M2 coefcients for each FF and FB lter, respectively. We dene the following vectors:
.u/ .u/ 1. x.u/0 [i] D [xk F0 [i]; : : : ; x k F0 Mi C1 [i]]T , signal samples stored at instant k F0 in the kF delay lines of the FF lters with input given by the demodulator output of user u at subchannel i; .u;v/ 2. c.u;v/ [i] D [c0 [i]; : : : ; c.u;v/1 [i]]T , coefcients of the FF lter from the demoduM1 lator output of user v to the decision element input of user u at subchannel i; .u;v/ 3. b.u;v/ [i] D [b1 [i]; : : : ; b.u;v/ [i]]T , coefcients of the FB lter from the decision M2 element output of user v to the decision element input of user u at subchannel i; 2 3T
k 1
D .u/[i] F0
instant k in the delay lines of the FB lters with input given by the decision element output of user u at subchannel i, where D .u/ [i] is a suitable integer delay related to .u/ the DFE; we assume no decision errors, that is ak [i] D ak [i]. O .u/ The input to the decision element of user u at subchannel i at instant k is then expressed by (see Section 8.5)
.u/ yk [i] D c.u;u/ [i] x.u/0 [i] C b.u;u/ [i] a.u/ [i] C kF k
T T
U X v6Du
vD1
(16.48)
.u/ l
D .u/ [i] F0
(16.49)
Without loss of generality, we extend the technique developed in Section 8.5 for the single-user fractionally-spaced DFE to determine the optimum coefcients of the DFE structure for user u D 1. We introduce the following vectors and matrices: 1. h.u/ [i] D G .u/ [i][h .u/ CM 1CD .u/ [i ] [i]; : : : ; h .u/ CD .u/ [i ] [i]]T , vector of M1 samples m m F0 m F0 1 of the impulse response of subchannel i of user u. 2. h.u;v/ [i] D G .v/ [i][h .u;v/ F CM 1CD .u;v/ [i ] [i]; : : : ; h .u;v/ F CD .u;v/ [i ] [i]]T , vecFEXT;m FEXT;m 0 FEXT;m 0 1 tor of M1 samples of the FEXT impulse response of subchannel i from user v to user u; assuming the differences between the propagation delays of the signal of user u and of the cross-talk signals originated by the other users are negligible, we have D .u;v/ [i] D D .u/ [i].
1134
M2 X mD1
V X vD2
! (16.50)
M2 X
2 2 .a .l/ [i ] h.l/ [i] h.l/ [i] C a .1/ [i ] h.l;1/ [i] h.l;1/ [i]/ FEXT;m FEXT;m m m
T T
mD1
(16.51)
mD1
M2 XB T B 2.1/ h.l;1/ [i] h.1/T [i] C 2.l/ h.l/ [i] h.l;1/ [i] FEXT;m m @ a [i ] FEXT;m a [i ] FEXT;m
1 C
U X
pD2
p6Dl
.l/
. j/T
[i]]
. j;1/T
.l; j/
. j/T
mD1
1 < j < l U (16.53) where all above matrices are M1 M1 square matrices. 4. R.1;1/ [i] R.1;2/ [i] 6 R.2;1/ [i] R.2;2/ [i] 6 R.1/ [i] D 6 : : : : 4 : : .U;1/ [i] R.U;2/ [i] R 2 3 : : : R.1;U / [i] : : : R.2;U / [i] 7 7 7 : :: : 5 : : .U;U / [i] ::: R
(16.54)
1135
where R.1/ [i] in general is a positive semi-denite Hermitian matrix, for which we assume here the inverse exists. 5.
.1/ .2;1/ .U;1/ 2 p.1/ [i] D a .1/ [i ] [h0 [i]; hFEXT;0 [i]; : : : ; hFEXT;0 [i]]T
T T T
(16.55)
Dening the vectors c.1/ [i] D [c.1;1/ [i]; c.1;2/ [i]; : : : ; c.1;U / [i]]T , and b.1/ [i] D .1;1/T [i]; : : : ; b.1;U /T [i]]T , the optimum coefcients are given by [b
.1/ copt [i] D [R.1/ [i]] 1 .1/
p [i] 3
(16.56)
and 2 6 7 6 7 6 7 6 7 6 7 6 7 6 7 T T T .U;1/ 6 h.1/ [i]; h.2;1/ 7 6 7 M1 FEXT;M1 [i]; : : : ; hFEXT;M1 [i] 6 7 6 7 .1;2/T .2/T .U;2/T 6 7 hFEXT;1 [i]; h1 [i]; : : : ; h1 [i] 6 7 6 7 : 6 7 6 7 : 6 7 6 7 .1/ .1/ : bopt [i] D 6 7 copt [i] T T T .2/ .U;2/ 6 h.1;2/ [i]; hFEXT;M1 [i]; : : : ; hFEXT;M1 [i] 7 6 FEXT;M1 7 6 7 : 6 7 6 7 : 6 7 6 7 : 6 7 6 7 T T T 6 7 h.1;U / [i]; h.2;U / [i]; : : : ; h.U / [i] 6 7 FEXT;1 FEXT;1 1 6 7 : 6 7 6 7 : 6 7 6 7 : 4 5 .1;U /T .2;U /T .U /T hFEXT;M1 [i]; hFEXT;M1 [i]; : : : ; hFEXT;M1 [i] The MMSE value at the decision point of user 1 on subchannel i is thus given by
.1/ 2 Jmin [i] D a .1/ [i ] .1/ p.1/ [i]copt [i]
H
(16.57)
(16.58)
The performance of an OFDM system is usually measured in terms of achievable bit rate for given channel and cross-talk characteristics (see Chapter 13). The number of bits per modulation interval than can be loaded with a bit-error probability of 10 7 on subchannel i is given by (see (13.15)) ! 2 a .1/ [i ] 10.G code 0 gap;d B /=10 b.1/ [i] D log2 1 C .1/ (16.59) Jmin [i]
1136
where G code is the coding gain assumed to be the same for all users and all subchannels. The achievable bit rate for user 1 is therefore given by
.1/ Rb D
(16.60)
16.4.1
Upstream power back-off (PBO) methods are devised to allow remote users in a VDSL system to achieve a fair distribution of the available capacity in the presence of FEXT [12]. The upstream VDSL transmission rates, which are achievable with PBO methods, usually depend on parameters, for example, a reference length or the integral of the algorithm of the received signal power spectral density, that are obtained as the result of various trade-offs between services to be offered and allowed maximum line length. However, the application of such PBO methods results in a suboptimum allocation of the signal power for upstream transmission. It is desirable to devise a PBO algorithm with the following characteristics: for each individual user, the transmit signal PSD is determined by taking into account the distribution of known target rates and estimated line lengths of users in the network, and the total power for upstream signals is kept to a minimum to reduce interference with other services in the same cable binder. In the preceding section, we found the expression (16.60) of the achievable upstream rate for a user in a VDSL system with U users, assuming perfect knowledge of FEXT impulse responses and multiuser detection. However, PBO may be applied by assuming only the knowledge of the statistical behavior of FEXT coupling functions with no attempt to cancel interference. In this case, the achievable bit rate of user u is given by 2 3
.u/ Rb
6 6 6 P .u/ . f /jH.u/ . f /j2 6 D log2 61 C 10.G code U 6 X B .u;v/ 6 P .v/ . f /jHFEXT . f /j2 C N0 4 Z
v6Du
vD1
(16.61)
where P .u/ . f / denote the PSD of the signal transmitted by user u, H.u/ . f / is the frequency .u;v/ response of the channel for user u, HFEXT . f / is the FEXT frequency response from user v to user u, and N0 is the PSD of additive white Gaussian noise. From Section 4.4.2, the expression of the average FEXT power coupling function is given by
.u;v/ jHFEXT . f /j2 D kt f 2 min.L u ; L v /jH.v/ . f /j2
(16.62)
where L u and L v denote the lengths of the lines of user u and v, respectively, and kt is a constant.
1137
Assuming that the various functions are constant within each subchannel band for OFDM modulation, we approximate (16.61) as 2 3
.u/
Rb D
M 1 X
i D0
6 6 6 1 P .u/ . f i /jH.u/ . f i /j2 6 10.G code log2 61 C U 6 T X .u;v/ 6 P .v/ . f i /jHFEXT . f i /j2 C N0 4
v6Du
vD1
where f i denotes the center frequency of subchannel i. .u/ Let P . f i / denote the PSD of the signal transmitted by user u on subchannel i with .u/ [i] D 1. Then the PBO problem can be formulated as follows: nd the minimum gain G of the function U U X XZ XM 1 1 .u/ .G .u/ [i]/2 P . f i / P .u/ . f / d f ' (16.64) T B uD1 uD1 i D0 subject to the constraints: 1. 0 .G .u/ [i]/2 P and 2.
.u/ .u/ Rb Rb;target .u/
. f i / Pmax
u D 1; : : : ; U
i D 0; : : : ; M
(16.65)
u D 1; : : : ; U
(16.66)
.u/ where Pmax is a constant maximum PSD value and Rb;target is the target rate for user u. .u/ In (16.66), Rb is given by (16.60) or (16.63), depending on the receiver implementation. Finding the optimum upstream transmit power distribution for each user is therefore equivalent to solving a non-linear programming problem in the U M parameters G .u/ [i], u D 1; : : : ; U , i D 0; : : : ; M 1. The optimum values of these parameters that minimize (16.64) can be found by simulated annealing [13, 14].
16.4.2
ETSI has dened two modes of operation, named A and B, for PBO [12]. For a scenario using upstream VDSL transmission of two adjacent links with unequal lengths, mode A states that the signal-to-noise ratio degradation to either link shall not exceed 3 dB relative to the equal-length FEXT case. Mode B requires that the signal-to-noise ratio on the longer line shall not be degraded relative to the equal-length FEXT case; furthermore, degradation to the signal-to-noise ratio on the shorter line shall be bounded such that the shorter line can support at least the upstream rate supported on the longer line. Several methods compliant
1138
with either mode A or B have been proposed. PBO methods are also classied into methods that allow shaping of the PSD of the transmitted upstream VDSL signal, e.g., the equalized FEXT method, and methods that lead to an essentially at PSD of the transmitted signal over each individual upstream band, e.g., the average log method. Both the equalized FEXT and the average log method, which are described below, comply with mode B. The equalized FEXT method requires that the PSD of user u be computed as [12] L ref jHref . f /j2 .u/ (16.67) Pmax ; Pmax P . f / D min L v jH.u/ . f /j2 where L ref and Href denote a reference length and a reference channel frequency response, respectively. The average log method requires that, for an upstream channel in the frequency band . f 1 ; f 2 /, user u adopt a constant PSD given by [15] P .u/ . f / D P. f 1 ; f 2 /
.u/
f 2 . f1; f2/
(16.68)
where P..u/; f 2 / is a constant PSD level chosen such that it satises the condition f1 Z
f2 f1
(16.69)
where K . f 1 ; f 2 / is a constant. In this section, the achievable rates of VDSL upstream transmission using the optimum algorithm (16.64) and the average log method are compared for various distances and services. The numerical results presented in this section are derived assuming a 26-gauge telephone twisted-pair cable (see Table 4.2). The noise models for the alien-cross-talk disturbers at the line termination and at the network termination are taken as specied in [16] for the ber-to-the-exchange case. Additive white Gaussian noise with a power spectral density of 140 dBm/Hz is assumed. We consider upstream VDSL transmission of U D 40 users over the frequency band given by the union of B1 = (2.9 MHz, 5.1 MHz) and B2 = (7.05 MHz, 12.0 MHz), similar to those specied in [12]. The maximum PSD value is Pmax D 60 dBm/Hz. FEXT power coupling functions are determined according to (16.62), where kt D 6:6510 21 . Upstream transmission is assumed to be based on FMT modulation with bandwidth of the individual subchannels equal to 276 kHz and excess bandwidth of 12.5%; for an efcient implementation, a frequency band of (0, 17.664 MHz) is assumed, with M D 64 subchannels, of which only 26 are used. For the computation of the achievable rates, for an error probability of 10 7 a signal-to-noise ratio gap to capacity equal to 0 0 gap;dB D 0 gap;dB C 6 D 15:8 dB, which includes a 6 dB margin against additional noise sources that may be found in the DSL environment [17], and G code D 5:5 dB are assumed. For each of the methods and for given target rates we consider two scenarios: the users are i) all the same distance L from the central ofce, and ii) uniformly distributed at ten different nodes, having distances j L max =10, j D 1; : : : ; 10, from the central ofce. To assess the performance of each method, the maximum line length L max is found, such that all users can reliably achieve a given target rate Rb;target D 13 MBit/s. The achievable rates are also computed for the case that all users are at the same distance from the central ofce and no PBO is applied.
1139
Figure 16.10. Achievable rates of individual users versus cable length using the optimum upstream PBO algorithm for a target rate of 13 Mbit/s. [From [10], c 2001 IEEE.]
For the optimum algorithm, the achievable rates are computed using (16.63). Furthermore, different subchannel gains may be chosen for the two bands, but transmission gains within each band are equal. Figure 16.10 shows the achievable rates for each group of four users with the optimum algorithm for the given target rate. The maximum line length L max for scenario ii) turns out to be 950 m. For application to scenario ii), the optimum algorithm requires the computation of 20 parameters. Note that for all users at the same distance from the central ofce, i.e., scenario i), the optimum algorithm requires the computation of two gains equal for all users. For scenario i), the achievable rate is equal to the target rate up to a certain characteristic length L max , which corresponds to the length for which the target rate is achieved without applying any PBO. Also note that L max for scenario ii) is larger than the characteristics length found for scenario i). Figure 16.11 illustrates the achievable rates with the average log algorithm (16.69). Joint optimization of the two parameters K B1 and K B2 for maximum reach under scenario ii) yields K B1 D 0:02 mW, K B2 D 0:05 mW, and L max D 780 m. By comparison with Figure 16.10, we note that for the VDSL transmission spectrum plan considered, optimum upstream PBO leads to an increase in the maximum reach of up to 20%. This increase depends on the distribution of target rates and line lengths of the users in the network. At this point, further observations can be made on the application of PBO. Equal upstream services have been assumed for all users. The optimum algorithm described is even better suited for mixed-service scenarios.
1140
Figure 16.11. Achievable rates of individual users versus cable length using the average log upstream PBO method for a target rate of 13 Mbit/s. [From [10], c 2001 IEEE.]
The application of PBO requires the transmit PSDs of the individual user signals to be recomputed at the central ofce whenever one or more users join the network or drop out of the network. To illustrate system performance achievable with multiuser detection, we consider U D 20 users, uniformly distributed at ten different nodes having distances j L max =10, j D 1; : : : ; 10, from the central ofce, where identication of FEXT impulse responses is performed. For the computation of the achievable rates of individual users, the FEXT impulse responses are generated by a statistical model, and L max D 500 m is assumed. Furthermore, to assess the relative merits of multiuser detection and coding, the achievable rates are computed for the two cases of uncoded and coded transmission. For coded transmission, a powerful coding technique yielding 8.5 dB coding gain for an error probability of 10 7 is assumed. Figure 16.12 illustrates the achievable rates for perfect suppression of all interferers, which corresponds to the single-user bound. For comparison, the achievable rates for the case that all users are the same distance away from the central ofce and neither multiuser detection nor coding are applied are also given. Figure 16.13 illustrates the achievable rates for perfect suppression of the ten worst interferers and no application of PBO. We observe that, without PBO, the partial application of multiuser detection does not lead to a signicant increase of achievable rates for all users, even assuming large coding gains. Finally, Figure 16.14 depicts the achievable rates obtained for perfect suppression of the ten worst interferers and application of the optimum PBO algorithm with target rate of Rb;target D 75 Mbit/s for all users. The target rate is achieved by all users with the joint application of multiuser detection, coding, and power back-off.
1141
Figure 16.12. Achievable rates of individual users versus cable length with all interferers suppressed. [From [10], c 2001 IEEE.]
Figure 16.13. Achievable rates of individual users versus cable length with ten interferers suppressed and no PBO applied. [From [10], c 2001 IEEE.]
1142
Figure 16.14. Achievable rates of individual users versus cable length with ten interferers suppressed and optimum PBO applied for a target rate of 75 Mbit/s. [From [10], c 2001 IEEE.]
To summarize the results of this section, a substantial increase in performance with respect to methods that do not require the identication of FEXT responses is achieved by resorting to reduced-complexity multiuser detection in conjunction with power back-off. This approach yields a performance close to the single-user bound shown in Figure 16.12.
Bibliography
[1] D. G. Messerschmitt and E. A. Lee, Digital communication. Boston, MA: Kluwer Academic Publishers, 2nd ed., 1994. [2] D. G. Messerschmitt, Design issues for the ISDN U-Interface transceiver, IEEE Journal on Selected Areas in Communications, vol. 4, pp. 12811293, Nov. 1986. [3] D. D. Falconer, Adaptive reference echo-cancellation, IEEE Trans. on Communications, vol. 30, pp. 20832094, Sept. 1982. [4] D. L. Duttweiler, Adaptive lter performance with nonlinearities in the correlation multiplier, IEEE Trans. on Acoustics, Speech and Signal Processing, vol. 30, pp. 578 586, Aug. 1982.
16. Bibliography
1143
[5] M. J. Smith, C. F. N. Cowan, and P. F. Adams, Nonlinear echo cancelers based on transposed distributed arithmetic, IEEE Trans. on Circuits and Systems, vol. 35, pp. 618, Jan. 1988. [6] G. Cherubini, Analysis of the convergence behavior of adaptive distributed-arithmetic echo cancelers, IEEE Trans. on Communications, vol. 41, pp. 17031714, Nov. 1993. [7] S. B. Weinstein, A passband data-driven echo-canceler for full-duplex transmission on two-wire circuits, IEEE Trans. on Communications, vol. 25, pp. 654666, July 1977. [8] M. Ho, J. M. Ciof, and J. A. C. Bingham, Discrete multitone echo cancellation, IEEE Trans. on Communications, vol. 44, pp. 817825, July 1996. [9] P. J. W. Melsa, R. C. Younce, and C. E. Rohrs, Impulse response shortening for discrete multitone transceivers, IEEE Trans. on Communications, vol. 44, pp. 1662 1672, Dec. 1996. [10] G. Cherubini, Optimum upstream power back-off and multiuser detection for VDSL, in Proc. GLOBECOM 01, San Antonio, TX, Nov. 2001. [11] A. Duel-Hallen, Decorrelating decision-feedback multiuser detector for synchronous code-division multiple-access channel, IEEE Trans. on Communications, vol. 41, pp. 285290, Feb. 1993. [12] Access transmission systems on metallic access cables; Very high speed Digital Subscriber Line (VDSL); Part 2: Transceiver specication, ETSI Technical Specication 101 270-2 V1.1.1, May 2000. [13] S. Kirkpatrik, C. D. Gelatt Jr., and M. P. Vecchi, Optimization by simulated annealing approach, Science, vol. 220, pp. 671680, May 1983. [14] D. Vanderbilt and S. Louie, A Monte Carlo simulated annealing approach to optimization over continuous variables, J. Comp. Phys., vol. 56, pp. 259271, 1984. [15] Constant average log: robust new power back-off method, Contribution D.815 (WP1/15), ITU-T SG 15, Question 4/15, Apr. 1999. [16] ETSI VDSL specications (Part 1) functional requirements, Contribution D.535 (WP1/15), ITU-T SG 15, Question 4/15. June 21 V- July 2, 1999. [17] T. Starr, J. M. Ciof, and P. J. Silverman, Digital subscriber line technology. Upper Saddle River, NJ: Prentice-Hall, 1999.
Chapter 17
Wired and wireless network technologies allow users to obtain services that are offered by various providers, for example, telephony, Internet access at a rate of several Megabit per second, on demand television programs, and telecommuting. Links make use of channels that allow full duplex transmission and exhibit sufcient capacity in the two directions of transmission. Wired network technologies make use of the existing cable infrastructure, consisting of 1) UTP cables originally laid in the customer service area (local loops) for access to the public switched telephone network (PSTN) over a telephone channel with band limited to about 4 kHz; 2) UTP cables installed in buildings or small geographic areas for the links between stations of a local-area network (LAN); and 3) coaxial cables currently used to distribute analog TV signals via cable TV networks. Wireless network technologies have gained widespread popularity among users; for example, technologies that allow the mobility of users are digital-enhanced cordless telecommunications (DECT) (see Chapter 18), personal access communications systems (PACS), and universal mobile telecommunications systems (UMTS). Wireless LANs have been developed to extend or replace wired LANs in environments where portable network stations are needed, allowing users to freely move around. The multichannel multipoint distribution service (MMDS) and the local multipoint distribution service (LMDS) are proposed as wireless cable networks that can offer high transmission capacity in geographical areas where the terrain does not present obstacles to the propagation of microwave radio signals. Hybrid structures are also considered, where the downstream transmission from a central ofce to the user and the upstream transmission in the opposite direction are performed using different transmission media, for example, services provided by direct broadcast satellite (DBS) in combination with upstream transmission over the PSTN.
17.1
17.1.1 Modem
The word modem is the contraction of mod ulator-demodulator. The function of a modem is to convert a data signal into an analog passband signal that can be transmitted over the
1146
Local office
Voiceband modem
endtoend PSTN
Figure 17.1. Illustration of a link between voiceband modems over the PSTN.
PSTN. A modem is dened full duplex if it can transmit and receive simultaneously over the same telephone channel, or half duplex in the other case (see Section 6.13.1 on page 522). A link between voiceband modems over the PSTN is illustrated in Figure 17.1. One of the technologies developed for data transmission over the PSTN is described by the ITU-T standard V.34. The V.34 modem uses a QAM scheme with 4-dimensional TCM (see Chapter 12) and exible precoding for transmission over channels with ISI (see Chapter 13). We recall that exible precoding permits achieving coding and shaping gains with minimum transmit power penalty for arbitrary constellations, provided that the transfer function of the overall discrete time system does not have zeros on the unit circle. Among the innovations introduced by the V.34 modem, we recall the initial probing of the channel, during which the passband is determined; based on the probing results, the symbol rate, which is in the range from 2400 to 3429 Baud, and the carrier frequency are dened. The information bit rate for the standard version V.34bis is in the range from 2.4 to 33.6 kbit/s; the selected value depends on the symbol rate and on the channel signal-to-noise ratio. The assumption that the link between two modems in the PSTN is completely analog does not lead to an accurate model for the majority of telephone channels available today; in fact the same network is essentially digital and transports signals that represent speech or data with a bit rate of 64 kbit/s. If modems used by service providers are linked to the PSTN by a digital channel that allows data transmission at a bit rate of 64 kbit/s, then only one analog local loop is found between the user and the rest of the network, as illustrated in Figure 17.2. Modems designed for the channel model just described are usually called PCM modems and can transmit data at a bit rate of 4056 kbit/s over channels with a frequency band that goes from 0 Hz to about 4 kHz (see Section 17.1.1). This technology is described by the ITU-T standard X.90. In Table 17.1 the characteristics of some modems are briey described. For full duplex modems frequency division duplexing (FDD) or echo cancellation (EC) is used. The column labelled ITU-T std identies the international standards dened by the International Telecommunications UnionTelecommunications Standardization Sector. The acronym TC stands for trellis coding (see Chapter 12); for a description of the various modulation techniques we refer to Chapter 6.
1147
DSP
8bit x 8 ks/s=64 kbit/s *** (service provider)
DSP DAC
*** Bchannels are not necessarily trans parent: e.g., robbedbit signalling (RBS) and digital attenuation (PAD) are possible
Figure 17.2. Illustration of a link between the PCM modem and the server modem. Table 17.1 Commercial modems.
Bit-rate Symbol Duplex (bit/s) rate (method) 300 300 full (FDD) 1200 1200 half 1200 600 full (FDD) 2400 1200 half 2400 600 full (FDD) 2400 1200 full (EC) 4800 1600 half 4800 2400 full (EC) 9600 2400 half 9600 2400 full (EC) 14400 2400 full (EC) 28800 3429 full (EC) 56000 <4000 full (EC)
ITU-T std V.21 V.23 V.22 V.26 V.22bis V.26ter V.27 V.32 V.29 V.32 V.32bis V.fast (V.34) X.90
Modulation 2-FSK 2-FSK 4-PSK 4-PSK 16-QAM 4-PSK 8-PSK 4-QPSK 16-AM/PM 32-QAM C TC 128-QAM C TC 1024-QAM C TC PAM C TC
(see Chapter 4), a reliable link over a local loop is preferable in many cases given the large number of already installed cables [1, 2, 3]. Figure 17.3 illustrates a link between two integrated services digital network (ISDN) modems for two rates: basic rate (BR) at 160 kbit/s and primary rate (PR) at 1.544 or 2.048 Mbit/s. The various digital subscriber line (DSL) technologies (in short xDSL) allow full duplex transmission between user and central ofce at bit rates that may be different in the two directions [4, 5]. For example, the high bit rate digital subscriber line (HDSL) offers a solution for full duplex transmission at a bit rate of 1.544 Mbit/s, also called T1 rate
1148
v Highdata rate DSL (HDSL): 4 or 2 wires, 12 kfeet [ 2 wires: singleline DSL (SDSL) ]
Figure 17.3. Illustration of links between ISDN modems on the subscriber line (BR D basic rate and PR D primary rate).
(see Section 6.13), over two twisted pairs and up to distances of 4500 m; the single-line high-speed DSL (SHDSL) provides full duplex transmission at rates up to 2.32 Mbit/s over a single twisted pair, up to distances of 2000 m. A third example is given by the asymmetric digital subscriber line (ADSL) technology (see Figure 17.4); originally ADSL was proposed for the transmission of video-on-demand signals; later it emerged as a technology capable of providing a large number of services. For example, ADSL-3 is designed for downstream transmission of four compressed video signals, each having a bit rate of 1.5 Mbit/s, in addition to the full duplex transmission of a signal with a bit rate of 384 kbit/s, a control signal with a bit rate of 16 kbit/s, and an analog telephone signal, up to distances of 3600 m. A further example is given by the very high-speed DSL (VDSL) technology, mainly designed for the ber-to-the-curb (FTTC) architecture. The considered data rates are up to 26 Mbit/s downstream i.e. from the central ofce or optical network unit to the remote terminal, and 4.8 Mbit/s upstream for asymmetric transmission, and up to 14 Mbit/s for symmetric transmission, up to distances not exceeding a few hundred meters. Figure 17.5 illustrates the FTTC architecture, where the link between the user and an optical network unit (ONU) is obtained by a UTP cable with maximum length of 300 m, and the link between the ONU and a local central ofce is realized by optical ber; in the gure, links between the user and the ONU that are realized by coaxial cable or optical ber are also indicated, as well as the direct link between the user and the local central ofce using optical ber with a ber-to-the-home (FTTH) architecture. Different baseband and passband modulation techniques are considered for high speed transmission over UTP cables in the customer service area. For example, the Study Group T1E1.4 of Committee T1 chose 2B1Q quaternary PAM modulation (see Example 6.5.1 on page 479) for HDSL, and DMT modulation (see Chapter 9) for ADSL. Among the organizations that deal with the standardization of DSL technologies we also mention the Study Group TM6 of the European Telecommunications Standard Institute (ETSI) and the Study Group 15 of the ITU-T. Table 17.2 summarizes the characteristics of DSL technologies; spectral allocations of the various signals are illustrated in Figure 17.6.
1149
ADSL1 modem
ADSL2 modem
12 kfeet = 3.6 km
Fiber
ONU
ADSL3 modem
ADSL3 modem
6 kfeet = 1.8 km
Figure 17.4. Illustration of links between ADSL modems over the subscriber line.
We now discuss more in detail the VDSL technology. Reliable and cost effective VDSL transmission at a few tens of Megabit per second is made possible by the use of frequencydivision duplexing (FDD) (see Section 6.13), which avoids signal disturbance by near-end cross-talk (NEXT), a particularly harmful form of interference at VDSL transmission frequencies. Ideally, using FDD, transmissions on neighboring pairs within a cable binder couple only through far-end cross-talk (FEXT) (see also Section 16.4), the level of which is signicantly below that of NEXT. In practice, however, other forms of signal coupling come into play because upstream and downstream transmissions are placed spectrally as close as possible to each other in order to avoid wasting useful spectrum. Closely packed transmission bands exacerbate interband interference by echo and NEXT from similar systems (self-NEXT), possibly leading to severe performance degradation. Fortunately, it is possible to design modulation schemes that make efcient use of the available spectrum and simultaneously achieve a sufcient degree of separation between transmissions in opposite directions by relying solely on digital signal processing techniques. This form of FDD is sometimes referred to as digital duplexing. The concept of divide and conquer has been used many times to facilitate the solution of very complex problems; therefore it appears unavoidable that digital duplexing for VDSL will be realized by the sophisticated version of this concept represented by multicarrier transmission, although single carrier methods have been also proposed. As discussed in
1150
Local office
ONU
VDSL modem
FTTC
300 m
Fiber
Cable modem
Coax
Cable modem
FTTH
Fiber Fiber modem
Acronym basic rate ISDN HDSL SHDSL ADSL ADSL lite VDSL
Standard
Bit rate (Mbit/s) 0.144 1.544, 2.048 0.1922.32 downstream 6:144 upstream 0:640 downstream 1:5 upstream 0:512 downstream 26 upstream 14
Distance (m) 6000 4000 2000 3600 best effort service 1500
1151
(a)
(b)
Figure 17.6. Spectral allocation of signals for xDSL technologies.
Section 9.5, various variants of multicarrier transmission exist. The digital duplexing method for VDSL known as Zipper [6] is based on discrete-multitone (DMT) modulation; here we consider ltered multitone (FMT) modulation (see Section 9.5), which involves a different set of trade-offs for achieving digital duplexing in VDSL and offers system as well as performance advantages over DMT [7, 8].
1152
The key advantages of FMT modulation for VDSL can be summarized as follows. First, there is exibility to adapt to a variety of spectrum plans for allocating bandwidth for upstream and downstream transmission by proper assignment of the subchannels. This feature is also provided by DMT modulation, but not as easily by single-carrier modulation systems. Second, FMT modulation allows a high-level of subchannel spectral containment and thereby avoids disturbance by echo and self-NEXT. Furthermore, disturbance by a narrowband interferer, e.g., from AM or HAM radio sources, does not affect neighboring subchannels as the side lobe lter characteristics are signicantly attenuated. Third, FMT modulation does not require synchronization of the transmissions at both ends of a link or at the binder level, as is sometimes needed for DMT modulation. As an example of system performance, we consider the bit rate achievable for different values of the length of a twisted pair, assuming symmetrical transmission at 22.08 MBaud and full duplex transmission by FDD based on FMT modulation. The channel model is obtained by considering a line with attenuation equal to 11.1 dB/100 m at 11.04 MHz (see Section 4.4.1), with 49 near-end cross-talk interference signals, 49 far-end cross-talk interference signals, and additive white Gaussian noise with a PSD of 140 dBm/Hz. The transmitted signal power is assumed equal to 10 dBm. The FMT system considered here employs the same linear-phase prototype lter for the realization of transmit and receive polyphase lter banks, designed for M D 256, K D 288, and D 10; we recall that with these values of M and K the excess bandwidth within each subchannel is equal to 12.5%. Per-subchannel equalization is obtained by a TomlinsonHarashima precoder (see Section 13.3.1) with 9 coefcients at the transmitter and a fractionally spaced linear equalizer with 26 T =2 spaced coefcients at the receiver. With these parameter values, and using the bit loading technique of Section 13.2, the system achieves bit rates of 24.9 Mbit/s, 10.3 Mbit/s, and 6.5 Mbit/s for the three lengths of 300 m, 1000 m, and 1400 m, respectively [9]. We refer to Section 16.4 for a description of the general case where users are connected at different distances from the central ofce and power control is applied.
17.1.2
High speed transmission over unshielded twisted pairs in local area networks
High speed transmission over UTP cables installed in buildings is studied by different standardization organizations. For example, the ATM Forum and the Study Group 13 of the ITU-T consider transmission at bit rates of 155.52 Mbit/s and above for the denition of the asynchronous transfer mode (ATM) interface between the user and the network. The IEEE 802.3 Working Group investigates the transmission at 1 Gbit/s over four twisted pairs, type UTP-5, for Ethernet (1000BASE-T) networks. UTP cables were classied by the EIA/TIA according to the transmission characteristics (see Chapter 4). We recall that UTP-3, or voice-grade, cables exhibit a signal attenuation and a cross-talk coupling much greater than that of UTP-5, or data-grade, cables. For LAN applications, the maximum cable length for a link between stations is 100 m. Existing cabling systems use bundles of twisted pairs, usually 4 or 25 pairs, and signals may cross line discontinuities, represented by connectors. For transmission over UTP-3 cables, in
1153
order to meet limits on emitted radiation, the signal band must be conned to frequencies below 30 MHz and sophisticated signal processing techniques are required to obtain reliable transmission. Standards for Ethernet networks that use the carrier sense multiple access with collision detection (CSMA/CD) protocol are specied by the IEEE 802.3 Working Group for different transmission media and bit rates. With the CSMA/CD protocol, a station can transmit a data packet only if no signal from other stations is being transmitted on the transmission medium. As the probability of collision between messages cannot be equal to zero because of the signal propagation delay, a transmitting station must continuously monitor the channel; in the case of a collision, it transmits a special signal called a jam signal to inform the other stations of the event, and then stops transmission. Retransmission takes place after a random delay. The 10BASE-T standard for operations at 10 Mbit/s over two unshielded twisted pairs of category 3 or higher denes one of the most widely used implementations of Ethernet networks; this standard considers conventional mono duplex (see Section 16.1) transmission, where each pair is utilized to transmit only in one direction using simple Manchester line coding (see Appendix 7.A) to transmit data packets, as shown in Figure 17.7. Transmitters are not active outside of the packets transmission intervals, except for transmission of a signal called link beat that is occasionally sent to assure the link connection. The request for transmission speeds higher than 10 Mbit/s motivated the IEEE 802.3 Working Group to dene standards for fast Ethernet that maintain the CSMA/ CD protocol and allow transmission at 100 Mbit/s and above. For example, the 100BASE-FX standard denes a physical layer (PHY) for Ethernet networks over optical bers. The 100BASE-TX standard instead considers conventional mono duplex transmission over two twisted pairs of category 5; the bit rate of 100 Mbit/s is obtained by transmission with a modulation rate of 125 MBaud and multilevel transmission (MLT-3) line coding combined with a channel code with rate 4/5 and scrambling, as illustrated in Figure 17.8. We also mention the 100BASE-T4 standard, which considers transmission over four twisted pairs of category 3; the bit rate of 100 Mbit/s is obtained in the following way by using an 8B6T code with a
Manchestercoded binary modulation or idle (no signal)
1 Idle 0 0 1 1 0 1 Idle
1/Tb = 10 MHz
1154
MLT3 coding
+2
0 1 0 1 0 2 0 1 0 0 1 +2
0 2 T = 8 ns Spectrum
0 (a) Medium independent interface (MII) Physical medium dependent (PMI) 125 Mbit/s MLT3 encoder Physical medium dependent (PMD)
62.5 MHz
Cat. 5
100 Mbit/s data or control info. Symbol encoder (rate4/5) Scrambling sequence generator
2pair UTP
MLT3 decoder
Figure 17.8. Illustration of (a) 100BASE-TX signal characteristics and (b) 100BASE-TX transceiver block diagram.
modulation rate equal to 25 MBaud: on the rst two pairs data are transmitted at a bit rate of 33.3 Mbit/s in half duplex fashion, while on the two remaining pairs data are transmitted at 33.3 Mbit/s in mono duplex fashion. A further version of fast Ethernet is represented by the 100BASE-T2 standard, which allows users of the 10BASE-T technology to increase the bit rate from 10 to 100 Mbit/s without modifying the cabling from category 3 to 5, or using four pairs for a link over UTP-3 cables. The bit rate of 100 Mbit/s is achieved with dual duplex transmission over two twisted pairs of category 3, where each pair is used to transmit in the two directions (see Figure 17.9). The 100BASE-T2 standard represents the most advanced technology for high
1155
Tx Tx pair 1
Tx Tx Rc pair 2 Rc
Rc Rc
Tb = 62.5 ns Spectrum The code violations (J,K) are used to mark the beginning and end of 802.5 frames. 0 1/T b = 16 MHz f
Figure 17.10. Illustration of signal characteristics for transmission over token ring networks.
speed transmission over UTP-3 cables in LANs; the transceiver design for 100BASE-T2 will be illustrated in Chapter 19. Other important examples of LANs are the token ring and the ber distributed data interface (FDDI) networks; standards for token ring networks are specied by the IEEE 802.5 Working Group, and standards for FDDI networks are specied by ANSI. In these networks the access protocol is based on the circulation of a token. A station is allowed to transmit a data packet only after having received the token; once the transmission has been completed, the token is passed to the next station. The IEEE 802.5 standard species operations at 16 Mbit/s over two unshielded twisted pairs of category 3 or higher, with mono duplex transmission. For the transmission of data packets Manchester line coding is adopted; the token is indicated by coding violations, as illustrated in Figure 17.10. The ANSI standard for FDDI networks species a physical layer for mono duplex transmission at 100 Mbit/s over two unshielded twisted pairs of category 5 identical to that adopted for the Ethernet 100BASE-TX standard. Table 17.3 summarizes the characteristics of existing standards for high speed transmission over UTP cables.
1156
Table 17.3 Scheme summarizing characteristics of standards for high speed transmission over UTP cables.
Standard
Bit rate (Mbit/s) 10 16 100 100 100 100 1000 100 25.6 51.84 100 155.52
Cable type
2-pair UTP-3 2-pair UTP-3 2-pair UTP-5 2-pair 4-pair 2-pair 4-pair UTP-5 UTP-3 UTP-3 UTP-5
Fiber Distributed Data Interface FDDI ANSI X3T9.5 Fast Ethernet 100BASE-TX 100BASE-T4 100BASE-T2 1000BASE-T4 AnyLAN 100 VG ATM User-Network Interfaces ATM-25 ATM-51 ATM-100 ATM-155 IEEE IEEE IEEE IEEE 802.3 802.3 802.3 802.3
IEEE 802.12 ATM ATM ATM ATM Forum Forum Forum Forum
4-pair UTP-3 2-pair 2-pair 2-pair 2-pair UTP-3 UTP-3 UTP-5 UTP-3
17.1.3
A hybrid ber/coax (HFC) network is a multiple-access network, in which a head-end controller (HC) broadcasts data and information for medium-access control (MAC) over a set of channels in the downstream direction to a certain number of stations, and these stations send information to the HC over a set of shared channels in the upstream direction [10, 11]. The topology of an HFC network is illustrated in Figure 17.11: an HFC network is a point-to-multipoint, tree and branch access network in the downlink, with downstream frequencies in the 50860 MHz band, and a multipoint-to-point, bus access network in the uplink, with upstream frequencies in the 542 MHz band. Examples of frequency allocations are shown in Figure 17.12. The maximum round-trip delay between the HC and a station is of the order of 1 ms. The IEEE 802.14 Working Group is one of the standardization bodies working on the specications for the PHY and MAC layers of HFC networks. A set of PHY and MAC layer specications adopted by North American cable operators is described in [12]. In the downstream direction transmission takes place in broadcast made over channels with bandwidth equal to 6 or 8 MHz, characterized by low distortion and high signal-to-noise ratio, typically 42 dB. The J.83 document of the ITU-T denes two QAM transmission schemes
1157
Headend controller
HC
Fiber node
FN
Trunk = Fiber
FN
Station
v 10 km
80 km 200000 km s)
5 40 54
650
750 MHz
for transmission in the downstream direction, with a bit rate in the range from 30 to 45 Mbit/s [13]; by these schemes spectral efciencies of 58 bit/s/Hz are therefore obtained. In the upstream direction the implementation of PHY and MAC layers is considerably more difcult than in the downstream. In fact, we can make the following observations: signals are transmitted in bursts from the stations to the HC; therefore it is necessary for the HC receiver to implement fast synchronization algorithms; signals from individual stations must be received by the HC at well-dened instants of arrival and power levels; therefore, procedures are required for the determination
1158
of the round-trip delay between the HC and each station, as well as for the control of the power of the signal transmitted by each station, as channel attenuation in the upstream direction may present considerable variations, of the order of 60 dB; the upstream channel is usually disturbed by impulse noise and narrowband interference signals; moreover, the distortion level is much higher than in the downstream channel. Interference signals in the upstream channel are caused by domestic appliances and HF radio stations; these signals accumulate along the paths from the stations to the HC and exhibit time-varying characteristics; they are usually called ingress noise. Because of the high level of disturbance signals, the spectral efciency of the upstream transmission is limited to 24 bit/s/Hz. The noise spectrum suggests that the upstream transmission is characterized by the possibility of changing the frequency band of the transmitted signal (frequency agility), and of selecting different modulation rates and spectral efciencies. In [12], a QAM scheme for upstream transmission that uses a 4 or 16 point constellation and a maximum modulation rate of 2.56 MBaud is dened; the carrier frequency, modulation rate and spectral efciency are selected by the HC and transmitted to the stations as MAC information.
1159
MAP
mini slots
CM tx opportunity
request contention
CM tx opportunity
maintenance
CM tx opp.
previous map
current map
future map
Figure 17.13. Example of a MAP message [12]. [Reproduced with permission of Cable Television Laboratories, Inc.]
maintenance interval is equivalent to the maximum round-trip delay plus the transmission time of a RNG-REQ message. At the instant specied in the MAP message, the station sends a rst RNG-REQ message using the lowest power level of the transmitted signal, and is identied by a SID equal to zero as a station that requests to join the network. If the station does not receive a response within a pre-established time, it means that a collision occurred between RNG-REQ messages sent by more than one station, or that the power level of the transmitted signal was too low; to reduce the probability of repeated collisions, a collision resolution protocol is used with random back-off. After the back-off time interval, of random duration, is elapsed, the station waits for a new MAP message containing an IE of initial maintenance and at the specied instant retransmits a RNG-REQ message with a higher power level of the transmitted signal. These steps are repeated until the HC detects a RNG-REQ message, from which it can determine the round-trip delay and the correction of the power level that the station must apply for future transmissions. In particular, the compensation for the round-trip delay is computed so that, once applied, the transmitted signals from the station arrive at the HC at well-dened time instants. Then the HC sends to the station, in a ranging response RNG-RSP message, the information on round-trip delay compensation and power level correction to be used for future transmissions; this message also includes a temporary SID. The station waits for a MAP message containing an IE of station maintenance, individually addressed to it by its temporary SID, and in turn responds through a RNG-REQ message, signing it with its temporary SID and using the specied round-trip delay compensation and power level correction; next, the HC sends another RNG-REQ message to the station with information for a further renement of round-trip delay compensation and power level correction. The steps of ranging request/response are repeated until the HC sends a ranging successful message; at this point the station can send a registration request
1160
(REG-REQ) message, to which the HC responds with a registration response (REG-RSP) message conrming the registration and specifying one or more SID that the cable modem must use during the following transmissions. Finite state machines for the ranging procedure and the regulation of the power level for the cable modem and for the HC are shown in Figures 17.14 and 17.15, respectively.
Time out T1
Time out T2
RNGRSP
Adjust local parameters Yes Retries exhausted ? No Adjust transmit power Time out T3 Random backoff Error reinitialize MAC RNGREQ Map with station maintenance opportunity
Time out T4
RNGRSP
Adjust local parameters Yes Retries exhausted ? No No Abort ranging set from HC ? Yes Wait for unicast maintenance opportunity Error reinitialize MAC Establish IP layer No Success set from HC ? Yes
Figure 17.14. Finite state machine used by the cable modem for the registration procedure [12]. [Reproduced with permission of Cable Television Laboratories, Inc.]
1161
RNGREQ
Yes
Increment retry counter in poll list for this CM Add CM to poll list for future maps
RNGRSP
RNGREQ
No
Retries exhausted ? Yes RNGRSP (abort ranging) Remove CM from poll list
Yes
No
Parameters within limits? Yes RNGRSP (success) Remove CM from poll list
Done
Figure 17.15. Finite state machine used by the HC for the registration procedure [12]. [Reproduced with permission of Cable Television Laboratories, Inc.]
17.2
Mobile radio technologies used to provide personal communication services to the end user are commonly grouped into high tier and low tier; the main characteristics of these
1162
Table 17.4 Scheme summarizing the main characteristics of high tier and low tier technologies.
High tier Base station Coverage/base station Bit rate Talk time for portables Vehicular speed Quality Principal application Typical examples Large and expensive Up to several km Low to medium Short (1 hour) >100 km/h <wireline Outdoor vehicular GSM, IS-95, UMTS, IMT-2000
Low tier Small and inexpensive <500 m radius Very high Long (>4 hours) 50 km/h wireline Indoor/outdoor pedestrian DECT, PACS, PHS, Bluetooth
Fixed part RJ 11
mobile
WLL applications
Figure 17.16. Illustration of the utilization of DECT in the wireless local loop.
technologies are summarized in Table 17.4. Appendix 17.A describes some of the widely adopted wireless technologies listed in Table 17.4. Low tier technologies are designed to achieve a quality of service (QoS) similar to that offered by wired networks. DECT in Europe and PACS in the United States were developed as technologies for the wireless local loop (WLL) to provide communication services also to mobile users [14, 15]. DECT was originally developed for application in wireless private branch exchanges (wireless PBXs) with low user mobility, and later extended as a technology for the WLL, as illustrated in Figure 17.16. PACS is illustrated in Figure 17.17.
17.2.1
For the transmission at bit rates of the order of 110 Mbit/s, wireless local area networks (WLANs) normally use the industrial, scientic, and medical (ISM) frequency bands dened by the United States Federal Communications Commission (FCC), that is 902928 MHz,
1163
Radio port RJ 11
Wireless access fixed unit Switch Radio port controller unit NI Network interface
mobile
WLL applications
Figure 17.17. Illustration of the utilization of PACS in the wireless local loop.
Mobile nodes
Backbone LAN
Router
Inbuilding LAN
24002483.5 MHz, and 57255850 MHz; these networks are also called RF WLANs to distinguish them from the IR WLANs that use the infrared (IR) frequency band. Specications for PHY and MAC layers of WLANs are developed by various standardization organizations, among which we cite the IEEE 802.11 Working Group and the European Telecommunications Standard Institute (ETSI). The region within which mobile stations have the possibility of exchanging information with the network is divided into cells (see Figure 17.24). To reduce interference, neighboring cells use different frequencies. Within each cell, an access station or base station allocates frequencies and guarantees to the mobile stations the possibility of accessing xed networks over metallic cables or optical bers (see Figure 17.18). In the United States WLANs are allowed to operate in the IMS frequency bands without needing a license from the FCC, which, however, sets restrictions on the power of the radio signal that must be less than 1 W and species that spread-spectrum technology (see Chapter 10) must be used whenever the signal power is larger than 50 mW. Most RF WLANs employ direct sequence or frequency hopping spread-spectrum systems;
1164
WLANs that use narrowband modulation systems usually operate in the band around 5.8 GHz, with a transmitted signal power lower than 50 mW, in compliance with FCC regulations. The coverage radius of RF WLANs is typically of the order of a few hundred meters. IR WLANs usually employ PAM-DSB modulation (see Appendix 7.C) with values of the carrier frequency of the order of the frequency of the infrared radiation, as also typically adopted in a ber optic link; this system allows a simple and low cost implementation of the PHY layer [16]. However, it is necessary to consider that non-negligible interference signals can be generated by infrared radiation sources, like the sun or lighting systems. Two methods are usually considered in designing IR WLANs. In the rst method the signal is transmitted along a well-dened direction; in this case an IR system can also be used outdoor with a coverage radius of a few kilometers. In the second method, the signal is irradiated in all directions and the coverage radius is limited to about 10 m.
Mobile station
CTS
DATA ACK
1165
The CSMA/CA principle is simple. All mobile stations that have packets to transmit compete for channel access by sending ready to transmit (RTS) messages by a CSMA protocol. If the base stations recognize a RTS message sent by a mobile station, it sends a clear to transmit (CTS) message to the same mobile station and this one transmits the packet; if reception of this packet occurs correctly, then the base station sends an acknowledgement (ACK) message to the mobile station. With CSMA/CA, the only possibility of collision will occur during the RTS stage of the protocol; however, we note that also the efciency of this protocol is reduced with respect to that of the simple CSMA/CD because of the presence of the RTS and CTS stages. As mobility is allowed, the number of mobile stations that are found inside a cell can change at any instant. Therefore it is necessary that each station informs the others of its presence as it moves around. A protocol used to solve this problem is the so-called hand-off, or hand-over, protocol, which can be described as follows: a switching station, or all base stations with a coordinated operation, registers the information relative to the signal levels of all mobile stations inside each cell; if a mobile station M is serviced by base station B1, but the signal level of station M becomes larger if received by another base station B2, the switching station proceeds to a hard hand-off operation whose nal result is that mobile station M is considered part of the cell covered by base station B2.
17.2.2
The multichannel multipoint distribution service is presently used in the United States to distribute TV signals for entertainment and education programming. It can offer a maximum number of 33 channels for analog TV in the band 2.1502.686 GHz, and the cell radius is typically in the range from 40 to 60 km, according to the environment characteristics and antenna positioning. The local multipoint distribution service is also considered for wireless access. The proposed spectrum allocation for this system comprises the bands 27.528.35 GHz and 29.129.25 GHz, and the cell radius is typically limited to 8 km. The conguration of MMDS and LMDS networks is illustrated in Figure 17.20. Assuming that most MMDS channels are dedicated to the distribution of TV signals, only a limited number of channels are available for downstream data transmission. An increase of system capacity can be obtained by methods of frequency reuse, for example cellularization or sectorization. Cellularization consists of subdividing the area into cells and designing the system so that there is a certain spatial separation between cells using the same frequencies. On the other hand, sectorization consists of using multiple directional antennas at the base station, where each antenna is used to distribute information to a certain group of users; frequency reuse means sending different signals to different user groups using the same RF channels. This method requires that the various user groups are located in geographical areas characterized by a sufcient azimuth separation and that antennas have a good rejection characteristic of the secondary lobes. An example of sectorization is illustrated in Figure 17.21, where six antennas are used; each antenna offers a 60-degree coverage, and to each antenna is assigned the same number of downstream channels. Moreover, two separate channel sets, labelled A and B,
1166
central office/
A1 B3 A3 B2 B1 A2
are considered; the two sets include the entire frequency band for downstream transmission and they are reused three times, thus tripling the system capacity with respect to the case without sectorization. The required rejection level of secondary lobes can be reduced by the combination of sectorization with orthogonal polarization of adjacent sectors; we note, however, that orthogonal polarization can be effectively introduced only in the case of signal propagation in the absence of multipath, as in point-to-point links over short distances, using at each end of the link antennas placed at a large height with respect
17. Bibliography
1167
to the ground. For residential area installations where roof top antennas are used, usually multipath components due to reected signals from the ground and surrounding buildings are present (see Section 4.6); the reected signals may have components with orthogonal polarity with respect to the desired signal and this phenomenon introduces interference. In the design of transmission systems for LMDS it is necessary to consider the fact that, even if the transmitter and receiver are in line of sight, the effect of trafc and foliage movement determines a transmission channel with very hostile fading characteristics; for example, it is common to encounter in these environments a Doppler spread of 2 Hz or less, with a fade of over 40 dB. To obtain a reliable transmission for LMDS and to overcome effects due to the multipath, in addition to sectorized directional antenna systems various techniques may be considered, such as: signal coding and adaptive equalization, frequency diversity, combining of received signals by multiple antenna systems.
Bibliography
[1] S. V. Ahamed, P. L. Gruber, and J.-J. Werner, Digital subscriber line (HDSL and ADSL) capacity of the outside loop plant, IEEE Journal on Selected Areas in Communications, vol. 13, pp. 15401549, Dec. 1995. [2] W. Y. Chen and D. L. Waring, Applicability of ADSL to support video dial tone in the copper loop, IEEE Communications Magazine, vol. 32, pp. 102109, May 1994. [3] G. T. Hawley, Systems considerations for the use of xDSL technology for data access, IEEE Communications Magazine, vol. 35, pp. 5660, Mar. 1997. [4] W. Y. Chen and D. L. Waring, Applicability of ADSL to support video dial tone in the copper loop, IEEE Communications Magazine, vol. 32, pp. 102109, May 1994. [5] T. Starr, J. M. Ciof, and P. J. Silverman, Understanding digital subscriber line technology. Englewood Cliffs, NJ: Prentice-Hall, 2000. [6] F. Sjoberg, M. Isaksson, R. Nilsson, P. Odling, S. K. Wilson, and P. O. Borjesson, Zipper: a duplex method for WDSL based on DMT, IEEE Trans. on Communications, vol. 47, pp. 12451252, Aug. 1999. [7] G. Cherubini, E. Eleftheriou, S. Olcer, and J. M. Ciof, Filter bank modulation techniques for very high-speed digital subscriber lines, IEEE Communications Magazine, vol. 38, pp. 98104, May 2000.
1168
R [8] G. Cherubini, E. Eleftheriou, and S. Olcer, Filtered multitone modulation for veryhigh-speed digital subscriber lines, IEEE Journal on Selected Areas in Communications, June 2002. R [9] G. Cherubini, E. Eleftheriou, and S. Olcer, Filtered multitone modulation for VDSL, in Proc. IEEE GLOBECOM 99, Rio de Janeiro, Brazil, pp. 11391144, Dec. 1999. [10] C. A. Eldering, N. Himayat, and F. M. Gardner, CATV return path characterization for reliable communications, IEEE Communications Magazine, vol. 33, pp. 6269, Aug. 1995. [11] C. Bisdikian, K. Maruyama, D. I. Seidman, and D. N. Serpanos, Cable access beyond the hype: on residential broadband data services over HFC networks, IEEE Communications Magazine, vol. 34, pp. 128135, Nov. 1996. [12] D. Fellows and D. Jones, DOCSIS(tm) , IEEE Communications Magazine, vol. 39, pp. 202209, Mar. 2001. [13] Digital multi-programme systems for television sound and data services for cable distribution, ITU-T Recommendation J.83, ITU-T Study Group 9, Oct. 24 1995. [14] W. Honcharenko, J. P. Kruys, D. Y. Lee, and N. J. Shah, Broadband wireless access, IEEE Communications Magazine, vol. 35, pp. 2027, Jan. 1997. [15] C. C. Yu, D. Morton, C. Stumpf, R. G. White, J. E. Wilkes, and M. Ulema, Lowtier wireless local loop radio systems. Part 1: Introduction, IEEE Communications Magazine, vol. 35, pp. 8492, Mar. 1997. [16] F. Gfeller and W. Hirt, A robust wireless infrared system with channel reciprocity, IEEE Communications Magazine, vol. 36, pp. 100106, Dec. 1998. [17] S. Singh, Wireless LANs, in The Mobile Communications Handbook (J. D. Gibson, ed.), ch. 34, pp. 540552, New York: CRC/IEEE Press, 1996. [18] R. O. LaMaire, A. Krishna, P. Bhagwat, and J. Panian, Wireless LANs and mobile networking: standards and future directions, IEEE Communications Magazine, vol. 34, pp. 8694, Aug. 1996. [19] M. Rahnema, Overview of the GSM system and protocol architecture, IEEE Communications Magazine, vol. 31, pp. 92100, Apr. 1993. [20] T. S. Rappaport, Wireless communications: principles and practice. Englewood Cliffs, NJ: Prentice-Hall, 1996. [21] D. J. Goodman, Wireless personal communication systems. Reading, MA: AddisonWesley, 1997. [22] R. Pirhonen, T. Rautava, and J. Penttinen, TDMA convergence for packet data services, IEEE Personal Communications Magazine, vol. 6, pp. 6873, June 1999.
17. Bibliography
1169
[23] A. Furuskar, S. Mazur, F. Muller, and H. Olofsson, EDGE: enhanced data rates for GSM and TDMA/136 evolution, IEEE Personal Communications Magazine, vol. 6, pp. 5666, 1999. [24] N. R. Sollenberger, N. Seshadri, and R. Cox, The evolution of IS-136 TDMA for third-generation wireless services, IEEE Personal Communications Magazine, vol. 6, pp. 818, 1999. [25] M. Austin, A. Buckley, C. Coursey, P. Hartman, R. Kobylinski, M. Majmundar, and J. P. Seymour, Service and system enhancements for TDMA digital cellular systems, IEEE Personal Communications Magazine, vol. 6, pp. 2033, 1999. [26] K. Pahlavan and A. H. Levesque, Wireless information networks. New York: John Wiley & Sons, 1995. [27] H. Holma and A. Toskala, eds, WCDMA for UMTS radio access for third generation mobile communications. Chichester: John Wiley & Sons, 2001. [28] B. A. Miller and C. Bisdikian, Bluetooth revealed. Upper Saddle River, NJ: PrenticeHall, 2001. [29] J. C. Haartsen and S. Mattison, Bluetooth a new lowpower radio interface providing shortrange connectivity, in Proc. of the IEEE, vol. 88, pp. 16511661, Oct. 2000. [30] T. Wilkinson, T. G. C. Phipps, and S. K. Barton, A report on HIPERLAN standardization, Int. Journal of Wireless Information Networks, vol. 2, pp. 99120, Apr. 1995. [31] R. van Nee, G. Awater, M. Morikura, and H. Takanashi, New high-rate wireless LAN standards, IEEE Communications Magazine, vol. 37, pp. 8288, Dec. 1999.
1170
Appendix 17.A
17.A.1
General observations
Before presenting some of the consolidated standards for wireless systems, we recall a few concepts that are of great interest in this eld.
Wireless systems
Although the trend, through third generation universal mobile telecommunications service (UMTS) systems, is to merge the different type of services into one large structure, a distinction must be made between two classes of wireless systems: voice-oriented systems and data-oriented systems. Figure 17.22 illustrates this aspect. From this scheme we note that both classes of systems can be further subdivided into two categories. Systems for voice transmission can be cordless, that is with low transmitted power and local area services [18], or cellular, that is with high transmitted power and wide area services. A similar distinction can be made for data transmission services: the wireless local area networks (WLANs) are low power systems with short range coverage, whereas the mobile-data networks are high power systems with long range coverage (see Sections 17.2 and 17.2.1).
1171
We will discuss here cellular systems, specically the standards GSM, IS-136, JDC, and IS-95; we will also address the standard DECT for cordless systems and HIPERLAN for WLANs.
Modulation techniques
The current second generation wireless systems are digital and use different types of modulation according to the required performance. For a discussion of modulation techniques for radio transmission systems, see Chapter 18. Using a QPSK (or a BPSK) system with a square root raised cosine transmit pulse yields a good trade-off between level of ISI and required bandwidth; in this case, however, the modulated signals do not have a constant envelope, which may represent a problem in the presence of a high power amplier. An alternative is given by GMSK, whose basic scheme is given in Figure 18.45; in this case the required bandwidth increases for higher values of Bt T , where Bt is the bandwidth of the Gaussian lter; the choice of the parameter Bt T is a compromise between pulse duration and system bandwidth. Figure 17.23 illustrates estimated power spectra of baseband QPSK and GMSK signals for two values of Bt T . The frequency is normalized by bit rate 1=Tb of the system; in this way it is easy to realize that the bandwidth efciency of QPSK is larger than that of GMSK.
Figure 17.23. Comparison between the PSD of QPSK and GMSK signals.
1172
1. Bandwidth efciency or spectral efciency: it is the ratio between the bit rate 1=Tb and the utilized bandwidth. 2. Power efciency: for the same 1=Tb and receiver complexity, the lower the transmitted power for a certain bit error probability, the greater the efciency. 3. Out-of-band radiation: it is the power outside of the main lobe of the power spectrum.1 4. Robustness to multipath: it expresses the insensitivity of the system to multipath channels. 5. Constant envelope: the use of power-efcient class C ampliers, because of their strong non-linearity, requires input signals with a constant envelope.
17.A.2
GSM standard
System characteristics
The global system for mobile communications (GSM) was started in the early 1990s with the aim of providing one standard within Europe. The services that this system offers are [19]:
1
For example, MSK has low bandwidth efciency and high out-of-band radiation; the Gaussian lter of a GMSK system attempts to mitigate both these problems.
1173
Figure 17.24. Illustration of cell and frequency reuse concepts; cells with the same letter are assigned the same carrier frequencies.
Telephony service: digital telephone service with guarantee of service to users that moves at a speed of up to 250 km/h; Data service: can realize the transfer of data packets with bit rates in the range from 300 to 9600 bit/s. ISDN service: some services, as the identication of a user that sends a call and the possibility of sending short messages (SMS), are realized by taking advantage of the integrated services digital network (ISDN), whose description can be found in [20]. A characteristic of the GSM system is the use of the subscriber identity module (SIM) card together with a four-digit number (ID); inserting the card in any mobile terminal, it identies the subscriber who wants to use the service. Important is also the protection of privacy offered to the subscribers of the system. Figure 17.25 represents the structure of a GSM system, that can be subdivided into three subsystems. The rst subsystem, composed of the set of BTSs and mobile terminals or mobile stations (MSs), is called a radio subsystem; it allows communication between the MSs and the mobile switching center (MSC), that coordinates the calls and
1174
also other system control operations. To a MSC are linked many base station controllers (BSCs); each BSC is linked up to several hundreds BTSs, each of which identies a cell and directly realizes the link with the mobile terminal. The hand-off procedure between two BTSs is assisted by the mobile terminal in the sense that it is a task of the MS to establish at any instant which BTS is sending the strongest signal. In the case of the hand-over between two BTS linked to the same BSC, the entire procedure is handled by the BSC itself and not by the MSC; in this way the MSC can save many operations. The second subsystem is the network switching subsystem (NSS), that in addition to the MSC includes: the home location register (HLR): this is a database that contains information regarding subscribers who reside in the same geographical area as the MSC; the visitor location register (VLR): this is a database that contains information regarding subscribers that are temporarily under the control of the MSC, but do not reside in the same geographical area;
1175
the authentication center (AUC): this controls codes and other information for correct communications management; the operation maintenance centers (OMC): they take care of the proper functioning of the various blocks of the structure. Finally, the MSC is directly linked to the public networks: PSTN for telephone services, ISDN for particular services as SMS, and data network for the transmission of data packets.
Radio subsystem
We now give some details with regard to the radio subsystem. The total bandwidth allocated for the system is 50 MHz; frequencies that go from 890 to 915 MHz are reserved for MS-BTS communications, whereas the bandwidth 935960 MHz is for communications in the opposite direction.2 In this way a full-duplex communication by frequency division duplexing (FDD) is realized. Within the total bandwidth there are 248 carriers allocated, that identify as many frequency channels called ARFCN; of these 124 are for uplink communications and 124 for downlink communications. The separation between two adjacent carriers is 200 kHz; the bandwidth subdivision is illustrated in Figure 17.26. Full-duplex communication is achieved by assigning two carriers to the user, one for transmission and one for reception, such that they are about 45 MHz apart. Each carrier is used for the transmission of an overall bit rate Rb of 270.833 kbit/s, corresponding to a bit period Tb D 3:692 s. The system employs GMSK modulation with parameter Bt T equal to 0.3; the aim is to have a power-efcient system. However, the bandwidth efciency is not very high; in fact we have 270:833 ' 1:354 bit/s/Hz (17.1) 200 which is smaller than that of other systems. Besides this FDM structure, there is also a TDMA structure; each transmission is divided into eight time intervals, or time slots, that identify the TDM frame. Figure 17.27 shows the structure of a frame as well as that of a single time slot. D
There also exists a version of the same system that operates at around the frequency of 1.8 GHz (in the USA the frequency is 1.9 GHz).
1176
Figure 17.27. TDM frame structure and slot structure of the GSM system.
As a slot is composed of 156.25 bits (not an integer number because of the guard time equal to 8.25 bits), its duration is about 576.92 s; therefore the frame duration is about 4.615 ms. In Figure 17.27 it is important to note the training sequence of 26 bits, used to analyze the channel by the MS or BS. The ag bits signal if the 114 information bits are for voice transmission or for control of the system. Finally, the tail bits indicate the beginning and the end of the frame bits. Although transmissions in the two directions occur over different carriers, to each communication is dedicated a pair of time slots spaced 4 slots apart (one for the transmit station and one for the receive station); for example, the rst and fth or the second and sixth, etc. Considering the sequence of 26 consecutive frames, that have a duration of about 120 ms, the 13th and 26th frames are used for control; then in 120 ms a subscriber can transmit (or receive) 114 24 D 2736 bits, which corresponds to a bit rate of 22.8 kbit/s. Indeed, the net bit rate of the message can be 2.4, 4.8, 9.6, or 13 kbit/s. Redundancy bits are introduced by the channel encoder for protection against errors, so that we get a bit rate of 22.8 kbit/s in any case. The original speech encoder chosen for the system was a RELP (see Chapter 5), improved by a long-term predictor (LTP), with a bit rate of 13 kbit/s. The use of a voice activity detector (VAD) allows an improvement in system capacity by reducing the bit rate to a minimum value during silence intervals in the speech signal. For channel coding, a convolutional encoder with code rate 1/2 is used. In [20] the most widely used speech and channel encoders are described, together with the data interleavers. In Figure 17.28 a scheme is given that summarizes the protection mechanism against errors used by the GSM system. The speech encoder generates, in 20 ms, 260 bits; as the
1177
bit rate per subscriber is 22.8 kbit/s, in 20 ms 456 bits must then be generated, introducing a few redundancy bits.3 To achieve reliable communications in the presence of multipath channels with delay spread up to 16 s, at the receiver equalization by a DFE and/or detection by the Viterbi algorithm are implemented. In [20, 21] the calling procedure followed by the GSM system and other specic details are fully described.
GSM-EDGE
Recently [22, 23], to support services with higher bit rates, the enhanced data for GSM evolution (EDGE) system was introduced, which employs 8-PSK, with a pulse represented in Figure 18.57, in place of GMSK. The bit rate is almost tripled to 69.2 kbit/s per subscriber; the channel encoder is also modied.
17.A.3
IS-136 standard
The standard IS-136 is an extension of the digital cellular IS-54 system; in turn IS-54 (called United States digital cellular (USDC)), was developed in the early 1990s as an extension of the analog standard AMPS, existing in the United States since the beginning of the 1980s. In many respects IS-136 is similar to GSM, in others is completely different.
3
The described slot structure and coding scheme refer to the transmission of user information, namely speech. Other types of communications, as for the control and management of the system, use different coding schemes and different time slot structures (see [21]).
1178
The structure of the system (see Figure 17.25) and types of services provided are similar to those of GSM; even IS-136 uses a combination of TDMA and FDMA, although with different specications. A rst substantial difference between the two systems is given by the modulation adopted: IS-136 uses a =4-DQPSK with a roll-off factor of 0.35. A considerable improvement in the bandwidth efciency is thus obtained. The bandwidth efciency is given by 48:6 ' 1:62 bit/s/Hz (17.2) D 30 where the numerator is given by the bit rate per carrier and the denominator by the spacing between carriers. The price for a better bandwidth efciency is a decrease of the power efciency with respect to the GSM. We examine some parameters used by the system. IS-136 has also a bandwidth of 50 MHz; the band 824849 MHz (or 18501865 MHz) is used for the MS-BS transmissions, whereas the band 869894 MHz (or 19301945 MHz) is reserved for communications in the reverse direction; again, the FDD technique is used to achieve full-duplex communication. Moreover, as in GSM, a combination of FDMA and TDMA is used to realize multiple access; the overall bandwidth is divided into sub-bands of 30 kHz, each carrying an overall bit rate of 48.6 kbit/s, that corresponds to a bit period of Tb ' 20:576 s. The structure of the TDM frame is illustrated in Figure 17.29, together with that of a single slot, which can be of two types, depending on whether it is used for communication from the mobile station to the base or vice versa. The frame is divided into 6 time slots, and each slot is composed of 324 bits, for a duration of about 6.667 ms, which means that the frame duration is of about 40 ms. As for GSM, for full-duplex communication two carriers, separated by about 45 MHz, are assigned: one for the MS-BS communication and the other for the reverse. This time, however, 2 time slots are assigned to the user, such that at most 3 users4 can use the same frame. We see from Figure 17.29 that, of the overall 324 bits per slot, only 260 bits are effective information, and the remaining serve as control and signaling for the system. Then in 40 ms there are only 260 2 D 520 information bits (transmitted or received), which corresponds to a bit rate of 13 kbit/s. The voice encoder is a vector-sum excitation linear predictive (VSELP) at 7.95 kbit/s, that generates only 159 bits in 20 ms (half a frame period). However, we have seen that in 20 ms there are 520=2 D 260 transmitted bits, hence the difference, that is 101 bits, is constituted by redundancy bits added by the channel encoder. Figure 17.30 illustrates channel coding for voice messages; for control messages the organization is different. The mechanism is similar to that of GSM; the 159 bits, generated by the voice encoder in 20 ms, are divided into two classes: the rst comprises the 77 most signicant bits (divided into a group of 65 bits and one of 12 bits), the second, the remaining 82 bits. After the procedure illustrated in the gure, the bits become 260, for a bit rate of 13 kbit/s. The type of demodulator/equalizer is not specied. For channels with a rms delay spread greater than 4:12 s, a RLS adaptive DFE was proposed [20]. For information on other characteristics of the IS-136 system and on its evolution, as well as on convergence with GSM-EDGE, we recommend references [24, 25].
4
In the GSM, 8 users share the same frame; in fact, to each user is assigned only one slot.
1179
Figure 17.29. TDM frame structure and slot structure for the IS-136 system.
Figure 17.30. Channel coding for voice messages of the IS-136 system.
1180
17.A.4
JDC standard
We discuss the Japanese digital cellular (JDC), which is very similar to the IS-136 standard. The JDC standard differs from the previous two standards mainly due to the fact that the total bandwidth of 80 MHz allocated for this system is much larger than that of GSM and IS-136. The frequency interval 940956 MHz is reserved for base-mobile station communications, whereas the band 810826 MHz is for communications in the opposite direction. To these two sub-bands two other sub-bands are added: the interval 14771501 MHz is for BS-MS communication, and the interval 14291453 MHz for MS-BS communications. The rst pair of sub-bands is called low band (LB), the second high band (HB). The FDD technique is used to achieve full-duplex communication; if the LB is used, the transmit carrier and the receive carrier are separated by 130 MHz; if we use the HB, the separation is only 48 MHz. Channels have a bandwidth of 25 MHz, both in the LB and in the HB; each channel is used for transmission of an overall bit rate Rb D 42 kbit/s, that corresponds to a bit period of about 23:81 s. Similar to the other two systems, besides the described FDMA structure, a TDMA structure is also used; a frame consists of 3 time slots and each one is composed of 280 bits (224 of information, 56 of control and signaling), which corresponds to a frame period of 3 280 23:81 D 20 ms. As one slot per frame is utilized, 280 bits (of which only 224 are of information) the user can transmit (or receive) information in 20 ms, which corresponds to a bit rate of 11.2 kbit/s. The speech encoder is a VSELP operating at 6.7 kbit/s and, after channel coding, the bit rate increases to 11.2 kbit/s. Channel coding is performed by a convolutional encoder with code rate 9/17. =4-DQPSK with a roll-off factor of 0.5 is adopted. As for the other two standards, the hand-off is assisted by the mobile terminal to reduce the MSC load.
17.A.5
IS-95 standard
As IS-136, the IS-95 system is a digital cellular system for North American geographical areas; it was also developed to be compatible with the AMPS standard, and is based on the CDMA spread spectrum technique, rather than on TDMA as the previously described systems [20, 26]. FDMA is still used: the total bandwidth allocated to the system is identical to that of IS-136 and has the same organization,5 but the channel bandwidth is 1.25 MHz. Also FDD is still used with a separation of 45 MHz between transmit and receive carriers. An important characteristic of this system is that the bit rate of the speech encoder can vary to lower power consumption, and is reduced to a minimum during silence intervals. The voice encoder is a Qualcomm codebook excited linear prediction (QCELP) with variable bit rate: 1.2, 2.4, 4.8, or 9.6 kbit/s.6 Interesting is the fact that the MS-BS communication greatly differs from the reverse: in particular, different sequences are employed to perform spreading. When the BS is
Within the total bandwidth, each of the two systems should use only a portion of the spectrum so that they do not interfere with each other. 6 In fact, the effective rate is from 1 to 8 kbit/s and the higher rates are due to the fact that additional redundancy bits have already been included.
1181
transmitting, data to all mobile terminals within the same cell (identied by a carrier) are sent using a different spreading sequence for each user. Regarding the MS, in transmission the communication occurs asynchronously, with the requirement that the power level of the received signals must be as uniform as possible. Whatever the bit rate used by the voice encoder, the channel encoder brings the bit rate to 19.2 kbit/s. The spreading procedure used for the BS-MS transmission is illustrated in Figure 17.31. Note that, after channel coding, interleaving is performed. Then the bit rate is increased to 1.2288 Mbit/s using one of 64 Walsh codes; to each user in a cell is assigned a different code, in order to maintain separation among signals of different users in the same cell and eliminate interference, at least in the absence of multipath. Moreover, to reduce the level of interference between two users that are in different cells, and to which the same Walsh code is assigned, scrambling is used; without going into details (see Chapter 10), the function of scrambling is to achieve quasi-orthogonality between the signals of the two users, also for different lags, so that the interference is kept at a low level [20]. Before the two modulation lters, a short pilot PN sequence is also inserted, that is used at the receiver for various purposes, e.g., channel identication. The system for the MS-BS communication has a different structure: channel coding is performed by a convolutional encoder with code rate 1/3, that yields a bit rate of 28.8 kbit/s. Spreading is still done by selecting one of 64 Walsh codes, but in a different way as compared to the BS-MS transmission [20]. QPSK is adopted; in particular, the modulation used by the mobile station is OQPSK. For detection, a RAKE receiver is used. We conclude this section by mentioning the third generation (3G) mobile radio system denoted universal mobile telecommunications system (UMTS), in the process of being standardized. 3G systems are intended for the transmission at rates from 16 kbit/s up to
Figure 17.31. Spreading procedure used by the IS-95 system for transmission from the base station to the mobile station.
1182
2 Mbit/s. Various approaches have been identied for the denition of the link layer; the most important is the wideband CDMA (WCDMA) system [27].
17.A.6
DECT standard
The digital European cordless telecommunications (DECT) system is a cordless digital system used in Europe. Unlike cellular systems that use cells with radius of the order of a few kilometers, the DECT system is mainly employed indoor and the cell radius is at most of a few tens of meters (typically 100 m). Services provided by the system are: speech transmission, data transmission, support to the ISDN network. The main characteristics of DECT are summarized in Table 17.5. For comparison, the main characteristics of PACS are summarized in Table 17.6. While originally the hand-off problem was not considered, as each MS corresponded to one BS only, now even for DECT we speak of hand-off assisted by the mobile terminal, such that the system conguration is similar7 to that of a cellular system (see Figure 17.25). An interesting characteristic is the use of the dynamic channel selection (DCS) algorithm, that allows the portable to know at every moment which channel (frequency) is the best (with the lowest level of interference) for communication and select it. We briey illustrate the calling procedure followed by the system: 1. When a mobile terminal wishes to make a call, it rst measures the received signals from the various BS8 and selects the one which yields the best signal level. 2. By the DCS algorithm, the mobile terminal selects the best free channels of the selected BS.
Table 17.5 Table summarizing the main characteristics of DECT.
Frequency range RF channel spacing Modulation Transmission bit rate Voice encoding method Access method Frame duration Subscriber TX peak power Radius of service Frequency planning
18801900 MHz 1728 kHz GMSK 1152 kbit/s 32 kbit/s ADPCM FDMA/TDMA/TDD 10 ms (24 time slots) 250 mW 100150 m Dynamic channel allocation
7 8
For a discussion of the differences between a cellular system and a cordless system, see Section 17.A.1. For DECT, the acronyms portable handset (PH) in place of MS and radio xed part (RFP) in place of BS, are often used.
1183
18501910 MHz uplink 19301990 MHz downlink RF channel spacing 300 kHz Modulation =4 QPSK Transmission bit rate 384 kbit/s Voice encoding method 32 kbit/s ADPCM Access method TDD/TDMA Frame duration 2.5 ms (8 time slots) Subscriber TX peak power 200 mW Radius of service 300500 m Frequency planning Quasi-static automatic frequency assignment
f1 f2 f3 f4 f5 f6 f7 f 8 f9 f 10
Frequency range
1900 MHz
3. MS sends a message, called access request, over the least interfered channel. 4. BS sends (or not) an answer: access granted . 5. If the MS receives this message, in turn it transmits the access conrm message and the communication starts. 6. If the MS does not receive the access granted signal on the selected channel, it abandons this channel and selects the second least interfered channel, repeating the procedure; after failing on 5 channels, the MS selects another BS and repeats all operations. The total band allocated to the system goes from 1880 to 1900 MHz and is subdivided into ten sub-bands, each with a width of 1.728 MHz; this is the FDMA structure of DECT represented in Figure 17.32. Each channel has an overall bit rate of 1.152 Mbit/s, that corresponds to a bit period of about 868 ns. Similar to the other systems, TDMA is used. The TDM frame, given in Figure 17.33, is composed of 24 slots: the rst 12 are used for the communication BS-MS, the other
1184
Figure 17.33. TDM frame structure and slot structure for the DECT system.
12 for the reverse communication; thus we realize a full-duplex communication by time division duplexing (TDD). In this DECT differs from all above considered wireless systems which use FDD; DECT allocates half of the frame for transmission and the other half for reception. In Figure 17.33 the slot structure is also shown; it is composed of 480 bits of which the rst 32 are xed and correspond to a synchronization word, the successive 388 are information bits, and the remaining 60 constitute the guard time. The frame has a duration of 480 24=1152 D 10 ms. The eld of 388 bits reserved for information bits is subdivided into two subelds A and B: the rst (64 bits) contains information for signaling and control of the system, the second (324 bits) contains user data. If the signal is speech, 4 of these 324 bits are parity bits, which translates into a net bit rate of 320 bits in a frame interval of 10 ms, and therefore in a bit rate of 32 kbit/s; an ADPCM voice encoder at 32 kbit/s is used and no channel coding is provided. For transmission, GMSK with parameter Bt T of 0.5 is adopted. At the MS receiver a pair of antennas are often used to realize switched antenna diversity. With this mechanism, fading and interference are mitigated. We summarize in Table 17.7 the characteristics of the various standards so far discussed. A standard that has in part overcome DECT is Bluetooth [28, 29], that operates around the frequency of 2:45 GHz, in the unlicensed and open industrial-scientic-medical (ISM) band. Bluetooth uses a FH/TDD scheme with Gaussian-shaped FSK (h 0:25). The modulation rate is 1 MBaud with a slot of 625 s; a different hop channel is used for each slot. This gives a nominal hop rate of 1600 hops/s. There are 79 hop carriers spaced by 1 MHz.
1185
Table 17.7 Summary of characteristics of the standards GSM, IS-136, JDC, IS-95, and DECT.
System Multiple access Band (MHz) Forward (BS!MS) Reverse (MS!BS) Duplexing Spacing between carriers Modulation Bit rate per carrier (kbit/s) Speech encoder bit rate (kbit/s) Convolutional encoder code rate Frame duration (ms)
GSM TDMA/ FDMA 935960 18051880 890915 17101785 FDD 200 kHz GMSK 0.3 270.833 RPE-LTP 13 1/2 4.615
IS-136 TDMA/ FDMA 869894 19301945 824849 18501865 FDD 30 kHz =4-DQPSK 48.6 VSELP 7.95 1/2 40
JDC TDMA/ FDMA 940960 14771501 810830 14291453 FDD 25 kHZ =4-DQPSK 42 VSELP 6.7
IS-95 FDMA/ CDMA 935960 824849 FDD 1250 kHz QPSK 1228.8 QCELP 1.29.6 1/2 (F) 1/3 (R) 20
DECT TDMA/ FDMA 18801900 18801900 TDD 1728 kHz GMSK 0.5 1152 ADPCM 32 none 10
20
All these standards use a CRC code, possibly together with a convolutional code.
The maximum bit rate is of 723:2 kbit/s in one direction and 57:6 kbit/s in the reverse direction.
17.A.7
HIPERLAN standard
In 1991 ETSI started the standardization process for the realization of an European high speed and low power data transmission system for WLAN applications. As cordless systems, also HIPERLAN supports mobile users, up to a maximum speed of about 35 km/h. HIPERLAN has a double application: the rst is that of replacing (see Figure 17.34) the last network link to the various users, in order to reduce costs due to the reconguration of the position of the terminal, the second is that of constituting, in indoor environments, a network among mobile (and xed) terminals. There are two versions of the HIPERLAN standard. HIPERLAN type 1. The allocated band for the system is from 5.15 to 5.30 GHz and from 17.1 to 17.3 GHz. The rst bandwidth of 150 MHz is still the most used and is further divided into 5 sub-bands; the rst carrier is at 5.176468 GHz, with a channel bandwidth of 23.5294 MHz. Figure 17.35 shows the band allocation.
1186
5300 MHz
The maximum overall (raw) bit rate is of 23.5294 Mbit/s with a corresponding bit interval of about 42.5 ns. The system is also able to work at lower bit rates (the so-called low rate of 1.4706 Mbit/s), in order to reduce power consumption. The transmission takes place by data packets: m blocks of 496 bits, with m that can vary from 1 to 47, are transmitted. Before the transmission of these blocks, there is the transmission of a training sequence of 450 bits, which is needed for channel identication and synchronization. Figure 17.36 shows the block diagram of the transmitter.
1187
Figure 17.36. Block diagram of the transmitter for the HIPERLAN type 1 system.
The source generates a sequence of m blocks, each one of 416 bits; at the output of the channel encoder the block length is of 496 bits. After interleaving, differential precoding is performed and successively a training sequence is inserted; the generated symbols are then input to the modulator. For HIPERLAN type 1, GMSK with parameter Bt T equal to 0.3 is adopted.9 At the receiver, adaptive equalization is performed before decoding [30]. HIPERLAN type 2. A more recent version of HIPERLAN allows transmission at a raw bit rate of 54 Mbit/s, for a net bit rate of up to 45 Mbit/s. The modulation is DMT-OFDM, with 48 subchannels. Other WLAN standards are described in [31], among which we mention the IEEE 802.11 standard.
Chapter 18
In this chapter, after an overview of front-end architectures for mobile radio receivers, we discuss modulation and demodulation schemes that are well suited for application to mobile radio systems because of their simplicity and robustness against disturbances introduced by the transmission channel. Appendix 17.A describes some of the standards where these schemes are adopted.
18.1
1190
Alternative architectures
In this section three receiver architectures are considered that attempt to integrate the largest number of receiver elements. To reach this objective, the A/D conversion of the scheme of Figure 18.1 must be shifted from baseband to IF or RF. In this case implementation of the channel selection by an analog lter bank is not efcient; indeed, it is more convenient to use a wideband analog lter followed by a digital lter bank. Reference is also made to multi-standard software-dened radio (SDR) receivers, or to the possibility of receiving signals that have different bandwidths and carriers, that are dened according to different standard systems. We can identify two approaches to the A/D conversion. 1. Full bandwidth digital conversion: by using a very high sampling frequency, the whole bandwidth of the SDR system is available in the digital domain, i.e. all desired channels are considered. Considering that this bandwidth can easily achieve 100 MHz, and taking into account the characteristics of interference signals, the dynamic of the ADC should exceed 100 dB. This solution, even though it is the most elegant, cannot easily be implemented as it presents a high complexity and high power consumption. 2. Partial bandwidth digital conversion: this approach uses a sampling frequency that is determined by the radio channel that presents the most extended bandwidth within the different systems we want to implement. The second approach will be considered, as it leads to an SDR system architecture that can be implemented with moderate complexity.
1191
towards the mixer input as well as towards the antenna, with consequent radiation. The interference signal generated by the oscillator could be reected on surrounding objects and be re-received: therefore this spurious signal would produce a time variant DC offset [1] at the mixer output. To understand the origin and consequences of this offset, we can make the following two observations. 1. The isolation between the oscillator input and the inputs of the mixer and linear amplier is not innite; in fact an LO leakage is determined by both capacitive coupling and device substrate. The spurious signal appears at the linear amplier and mixer inputs and is then mixed with the signal generated by the oscillator, creating a DC component; this phenomenon is called self-mixing. A similar effect is present when there is a strong signal coming from the linear amplier or from the mixer that couples with the LO input: this signal would then be multiplied by itself. 2. In Figure 18.2 to amplify the input signal, that is of the order of microvolts, to a level such that it can be digitized by a low cost and low power ADC, the total gain, from the antenna to the LPF output, is about 100 dB. Of this gain, 30 dB are usually provided by the linear amplier/mixer combination. With these data we can make a rst computation of the offset due to self-mixing. We assume that the oscillator generates a signal with a peak-to-peak value of 0.63 V and undergoes an attenuation of 60 dB when it couples with the LNA input. If the gain of the linear amplier/mixer combination is 30 dB, then the offset produced at the mixer output is of the order of 10 mV; if directly amplied with a gain of 70 dB, the voltage offset would saturate the circuits that follow. The problem is even worse if self-mixing is time variant. This event, as previously mentioned, occurs if the oscillator signal leaks to the antenna, thus being irradiated and reected back to the receiver from moving objects. Finally, the direct conversion receiver needs a tuner for the channel frequency selection that works at high frequency and with low phase noise; this is hardly obtainable with an integrated VCO with low Q.
1192
conversion, it utilizes a single mixer stage; the principal difference is that the frequency shift is not made to DC but to a small intermediate frequency. The main advantage of the low-IF system is that the desired channel has no DC components; therefore, the usual problems that emerge from DC offset present in the direct conversion are avoided.
18.2
18.2.1
In M-PSK, the transmitted signal (isolated pulse) is given by (6.128). For a continuous M-DPSK transmission with symbol period T , the transmitted signal is given by s.t/ D
C1 X kD 1
Re[e j
h T x .t
kT /e j2 f 0 t ]
(18.1)
where k is the phase associated with the transmitted symbol at instant kT given by the recursive equation (6.157).
1193
Figure 18.4. Non-coherent baseband differential receiver. Thresholds are set at .2n n D 1; 2; : : : ; M.
1/=M,
At the receiver the signal r is a version of s, ltered by the transmission channel and corrupted by additive noise. We denote as g A the cascade of passband lters used to amplify the desired signal and partially remove noise. As shown in Figure 18.4, let x be the passband received signal, centered around the frequency f 0 , equal to the carrier of the transmitted signal x.t/ D Re[x .bb/ .t/e j2 f 0 t ] (18.2)
where x .bb/ is the complex envelope of x with respect to f 0 . Using the polar notation x .bb/ .t/ D Mx .t/e j1'x .t/ , (18.2) can be written as x.t/ D Mx .t/ cos.2 f 0 t C 1'x .t// (18.3)
where 1'x .t/ is the instantaneous phase deviation of x with respect to the carrier phase (see Chapter 1, equation (1.207)). In the ideal case of absence of distortion and noise, sampling at suitable instants yields 1'x .t0 C kT / D k . Then for the recovery of the phase k we can use the receiver scheme of Figure 18.4, which, based on signal x, determines the baseband component y as y.t/ D y I .t/ C j y Q .t/ D 1 x .bb/ g Rc .t/ 2 (18.4)
The phase variation of the sampled signal yk D y.t0 C kT / between two consecutive symbol instants is obtained by means of the signal
z k D yk yk 1
(18.5)
Always in the ideal case and assuming that g Rc does not distort the phase of x .bb/ , z k turns out to be proportional to e jk . The simplest data detector is the threshold detector based on
1194
Note that a possible phase offset 1'0 and a frequency offset 1 f 0 , introduced by the receive mixer, yields a signal y given by y.t/ D [ 1 x .bb/ .t/e 2
j .1'0 C2 1 f 0 t/
] g Rc .t/
(18.7)
Assuming that g Rc does not distort the phase of x .bb/ , the signal vk becomes vk D 1'x .t0 C kT / 1'x .t0 C .k 1/T / 2 1 f 0 T (18.8)
which shows that the phase offset 1'0 does not inuence vk , while a frequency offset must be compensated by the data detector, summing the constant phase 2 1 f 0 T . The baseband equivalent scheme of the baseband differential receiver is given in Figure 18.5. The choice of h T x , g .bb/ , and g Rc is governed by the same considerations as in the A case of a QAM system; for an ideal channel, the convolution of these elements must be a Nyquist pulse.
18.2.2
The scheme of Figure 18.4 is introduced only to illustrate the basic principle, as its implementation complexity is similar to that of a coherent scheme. Indeed, this scheme has a reduced complexity because specications on the carrier recovery can be less stringent; moreover, it does not need phase recovery. An alternative scheme that does not use carrier recovery is illustrated in Figure 18.6. In this case the signal is rst delayed by a symbol period T and then multiplied by itself (I branch) and with its =2 phase-shifted version (Q branch) by a Hilbert lter. On the two branches the signals are given by I : x.t/x.t T / D Mx .t/ cos[2 f 0 t C 1'x .t/]Mx .t Q : x .h/ .t/x.t T / D Mx .t/ sin[2 f 0 t C 1'x .t/]Mx .t
T / cos[2 f 0 .t T / cos[2 f 0 .t
T / C 1'x .t T / C 1'x .t
T /] (18.9) T /]
1195
The lter g Rc removes the components around 2 f 0 ; the sampled lter outputs are then given by I : z k;I D Mx .t0 C kT /Mx .t0 C .k
1 2
If f 0 T D n, n an integer, or by removing this phase offset by phase shifting x or z k , it results z k;Q vk D tan 1 D 1'x .t0 C kT / 1'x .t0 C .k 1/T / (18.11) z k;I as in (18.6). The baseband equivalent scheme is shown in Figure 18.7, where, assuming that g Rc does not distort the desired signal, z k D 1 x .bb/ .t0 C kT /x .bb/ .t0 C .k 2 1/T / (18.12)
Typically, in this modulation system, the transmit lter h T x is a rectangular pulse or a Nyquist pulse; instead, g A is a narrow band lter to eliminate out of band noise.
1196
For a simple DBPSK, with k 2 f0; g, we only consider the I branch, and vk D z k;I is compared with a threshold set to 0.
Performance of M-DPSK
With reference to the transmitted signal (18.1), we consider the isolated pulse k s.t/ D Re[e j where
k
k
h T x .t
kT /e j2 f 0 t ]
(18.13)
k 1
C k
2.M 2 ;:::; k 2 0; M M
1/
(18.14)
In general, referring to the scheme of Figure 18.7, the ltered signal is given by x .bb/ .t/ D u .bb/ .t/ C w R .t/ where u .bb/ is the desired signal at the demodulator input (isolated pulse k) u .bb/ .t/ D e j
k
.bb/
(18.15)
kT /
(18.16)
.bb/ as .1=2/gCh is the complex envelope of the impulse response of the transmission channel, .bb/ and w R .t/ is zero mean additive complex Gaussian noise with variance 2 D 2N0 Brn , where Z C1 Z C1 2 jG A . f /j d f D jg A .t/j2 dt Brn D 1 1
Z D
C1 1
Z 1 .bb/ 2 G A . f / d f D 4
C1 1
1 .bb/ 2 g .t/ dt 4 A
(18.17)
is the equivalent bilateral noise bandwidth. We assume the following conguration: the transmit lter impulse response is given by r 2E s t T =2 h T x .t/ D rect (18.18) T T the channel introduces only a phase offset
.bb/ gCh .t/ D 2e j'a .t/
(18.19)
and the receive lter is matched to the transmit lter, r t T =2 2 .bb/ g A .t/ D rect T T At the sampling intervals t0 C kT D T C kT , let wk D w.bb/ .t0 C kT / R
(18.20)
(18.21)
1197
and r AD then x .bb/ .t0 C kT / D Ae j . with E[jwk j2 ] D 2 D 2N0 Brn D N0 since Brn D 1=2. Moreover, it results in z k D 1 x .bb/ .T C kT /x .bb/ .T C .k 2 D 1 [Ae j . k C'a / C wk ] [Ae j . k 2 1 D A | {z } C e j . k C'a / wk Ae jk 2 |
desired term
k C'a /
2E s 1 T 2
p 2 T D Es T
(18.22)
C wk
(18.23)
(18.24)
1/T /
1 C'a /
C wk 1 ] j.
k 1 C'a /
Ce
wk C
wk wk
(18.25)
disturbance
{z
A }
The desired p is similar to that obtained in the coherent case, M-phases on a circle term of radius A D E s . The variance of wk is equal to N0 , and I2 D N0 =2. However, even ignoring the term wk wk 1 =A, if wk and wk 1 are statistically independent it results in E[je j .
k C'a /
wk
Ce
j.
k 1 C'a /
wk j2 ] D 2 2 D 2N0
(18.26)
There is an asymptotic penalty, that is for E s =N0 ! 1, of 3 dB with respect to the coherent receiver case. Indeed, for a 4-DPSK, a more accurate analysis demonstrates that the penalty is only 2.3 dB for higher values of E s =N0 (see Section 6.5.1).
18.2.3
The scheme of Figure 18.8 makes use of a limiter discriminator (LD), as for frequency modulated (FM) signals, followed by an integrator lter over a symbol period, or integrate and dump (I&D) lter. Ideally, the discriminator output provides the instantaneous frequency deviation of x, i.e. 1 f .t/ D d P 1'x .t/ D 1'x .t/ dt (18.27)
1 f . / d D 1'x .t0 C kT /
1'x .t0 C .k
1/T / C 2n
(18.28)
1198
I m[.]
f(t)
t 0 +kT I&D
modulo 2
vk
Figure 18.10. Baseband equivalent scheme of a FM discriminator followed by an integrate & dump lter.
that coincides with (18.11) taking mod 2 . An implementation of the limiter-discriminator is given in Figure 18.9, while the baseband equivalent scheme is given in Figure 18.10, which employs the general relation Im x .bb/ .t/x .bb/ .t/ P 1 f .t/ D (18.29) 2 jx .bb/ .t/j2 In conclusion, all three schemes (baseband differential detector, differential detector and FM discriminator) yield the same output.
18.3
18.3.1 QPSK
Variants of QPSK
Basic schemes
From (18.1), we write the expression of the QPSK modulated signal C1 i h X s.t/ D Re e jk h T x .t kT /e j2 f 0 t e j'0
kD 1
(18.30)
1199
where k is the phase associated with the transmitted symbol at instant kT , and '0 is the carrier phase. We denote as ak the generic symbol at instant kT ak D e jk As k 2 f=4; 3=4g, we get 1 (18.32) ak 2 p f1; jg 2 For a modulation interval T D 2Tb , the map between bits and symbols is reproduced in Figure 18.11, with balanced symbols, b` 2 f 1; 1g. In particular, the following bit map is adopted, Re [ak ] D b2k
1
(18.31)
(18.33) (18.34)
b2k
1 h T x .t
kT /
(18.35)
b2k h T x .t
kT /
(18.36)
(18.37)
Q -1 1 1 1 ak
b 2k-1 1 -1
1200
+
0
b =1 1 b =1
0
b1=1 b =1
2
/2
b3=1 b =1
4
b5=1 b =1
6
s(t)
2T t
3T
4T
=4.
From Figure 18.11 we note that at successive instants, the phase fk g can undergo variations even equal to ; this implies large discontinuities in s with a consequent very high peak-power/average-power ratio of s if h T x is narrow band. In radio systems this can create saturation problems of the transmit amplier (see Section 4.8). Figure 18.12 shows the possible behavior of s for a wideband modulation pulse h T x .t/ D A rect for which s.t/ D A cos.2 f 0 t C '0 C k / kT t < .k C 1/T (18.39) t T =2 T (18.38)
b2k
1hT x t
1 T 2
(18.40)
Q.t/ D
b2k h T x .t
kT /
(18.41)
1201
Q -1 1 1 1 al
b 2k-1 1 -1
and as usual we have s.t/ D I .t/ cos.2 f 0 t C '0 / Q.t/ sin.2 f 0 t C '0 / (18.42)
In other words, the signal phase is varied alternatively, over the two branches, every Tb seconds; therefore the phase variation is now at most =2, as illustrated in Figure 18.13. For h T x as given by (18.38), we now have s.t/ D A cos.2 f 0 t C '0 C ` / for `Tb t < .` C 1/Tb (18.43)
For the same binary sequence of Figure 18.12, we show in Figure 18.14 the behavior of s given by (18.43). The reduced phase variation in s implies a greater compatibility with nonlinear ampliers. On the other hand, an OQPSK system requires a coherent demodulator.
(18.44)
1 C k
3 k 2 0; ; ; 2 2
(18.45)
The information is mapped in the phase variation between two successive symbols. At the receiver a non-coherent differential demodulator (see Section 18.2) is commonly employed. If we denote as e jk the information at instant kT , for the transmitted symbol ak the recursive relation ak D ak 1 e jk holds, from which we again obtain e jk by the equation e jk D ak ak 1 .
1202
0+k
0 b1=1
A
0 b0=1
/2 b1=1
b2=1
/2 b3=1
/2 b4=1
/2 b5=1
0 b6=1
s(t)
Tb
Tb
2Tb
3Tb t
4Tb
5Tb
6Tb
7Tb
=4.
/4-DQPSK
The transmitted symbol is ak D e j
k
, with
k
k 1
C k
(18.46)
where now k 2 f=4; 3=4; 5=4; 7=4g. Possible values of phase k are given in Figure 18.15. Note that for k even, k assumes values in f0; =2; ; 3=2g, and for k odd k 2 f=4; 3 =4; 5=4; 7=4g. Phase variations between two consecutive instants are =4 and 3=4, with a good peak/average power ratio of the modulated signal.
1203
18.3.2
Implementations
/4-DQPSK modulators
Due to its wide use, we examine in detail modulation and demodulation schemes for =4DQPSK. The modulator is similar to that for QPSK of Figure 18.16, differing only in the bit map. Now, for ak D ak;I C jak;Q from (18.46) a recursive equation for ak;I and ak;Q is given by ak;I D ak ak;Q D ak
1;I 1;Q
cos k
ak
1;Q 1;I
(18.48) (18.49)
cos k C ak
1204
Figure 18.17. Coherent demodulator for QPSK and OQPSK (CR = carrier recovery, STR = symbol timing recovery).
As k 2 f=4; 3=4; 5=4; 7=4g, then 1C j 1C j jk e 2 p ; p ; 2 2 and 8 > ak;I ; ak;Q 2 f0; 1g < 1 > ak;I ; ak;Q 2 p : 2
1 j 1 j p ; p 2 2 k even k odd
(18.50)
(18.51)
To demodulate a =4-DQPSK signal we can use one of the three differential schemes described in Section 18.2 or the coherent scheme illustrated in Figure 18.17. In any case, for an ideal AWGN channel performance of a =4-DQPSK is the same as for QPSK using a coherent receiver (see (6.153)) with differential encoding or a differential receiver (see (6.164)). For a 1BDD, for an ideal channel and using as transmit lter h T x and as receive lter g .bb/ a square root raised cosine pulse with roll-off D 0:3 ( D 1), Figure 18.18 A (Figure 18.19) illustrates the eye diagram at the decision point. We note that at the instant of maximum eye aperture the signal assumes the four possible values of e jk given by (18.50). Figures 18.20 and 18.21 illustrate eye diagrams for a LDI demodulator that are obtained using the same lters as in the previous case. As indicated by (18.28), at the instant of maximum eye aperture, the signal assumes the four possible values of k given by (18.46).
1205
Figure 18.18. Eye diagram of the 1BDD for =4-DQPSK, for an ideal channel and square root raised cosine receive lter with D 0:3.
Figure 18.19. Eye diagram of the 1BDD for =4-DQPSK, for an ideal channel and square root raised cosine receive lter with D 1.
1206
0.8
0.6
0.4
0.2
0.2
0.4
0.6
0.8
1 1
0.8
0.6
0.4
0.2
0 t/T
0.2
0.4
0.6
0.8
Figure 18.20. Eye diagram of the LDI demodulator for =4-DQPSK, for an ideal channel and square root raised cosine receive lter with D 0:3.
0.8
0.6
0.4
0.2
0.2
0.4
0.6
0.8
1 1
0.8
0.6
0.4
0.2
0 t/T
0.2
0.4
0.6
0.8
Figure 18.21. Eye diagram of the LDI demodulator for =4-DQPSK, for an ideal channel and square root raised cosine receive lter with D 1.
1207
18.4
The main advantage of FSK modulators consists in generating a signal having a constant envelope, therefore the distortion introduced by a HPA is usually negligible. However, they present two drawbacks: wider spectrum of the modulated signal as compared to amplitude and/or phase modulated systems, complexity of the optimum receiver in the presence of non-ideal channels. As already expressed in (6.72), a binary FSK modulator maps the information bits in frequency deviations ( f d ) around a carrier with frequency f 0 ; the possible transmitted waveforms are then given by s1 .t/ D A cos.2. f 0 f d /t C '1 / kT < t < .k C 1/T
(18.52)
p where, if we denote by E s the average energy of an isolated pulse, A D 2E s =T . Figure 18.22 illustrates the generation of the above signals by two oscillators, with f d and f 2 D f 0 C f d , selected at instants kT by the variable ak 2 frequency f 1 D f 0 f1; 2g related to the information bits. A particular realization of s1 and s2 is shown in Figure 18.23a. The resultant signal is given by s.t/ D
C1 X kD 1
sak wT .t
kT /
ak 2 f1; 2g
(18.53)
18.4.1
Following the derivation in [3], we consider the two cases of non-coherent and coherent FSK.
Acos(2 f2 t+ 2 ) ~ ak ~ s(t)
Acos(2 f1 t+ 1 )
1208
Figure 18.23. Binary FSK waveforms and transmitted signal for a particular sequence fak g.
C e j .2 f d tC'2 / ]
C1 X kD 1
wT .t
C1 X kD 1
kT / (18.55)
A C [e 2 We note that
j .2 f d t '1 /
e j .2 f d tC'2 / ]
ak wT .t
kT /
C1 X kD 1
wT .t
kT / D 1
(18.56)
while the second term of (18.55) is a PAM signal; using the results of Example 7.1.1 on page 544 it is easy to derive the continuous and the discrete parts of the power spectrum of s .bb/ .
1209
If WT . f / D T sinc. f T /e
j f T ,
we get
A2 N .c/ f d /j2 [jWT . f C f d /j2 C jWT . f Ps .bb/ . f / D 4T 2 .h/Re[e j .'2 '1 / WT . f C f d /W . f T A2 N .d/ [. f C f d / C . f Ps .bb/ . f / D f d /] 4 In (18.57), h D 2 f d T is the modulation index, and 1 h an integer .h/ D 0 otherwise
f d /]]
(18.57)
(18.58)
It can be seen that the power spectrum (18.57) contains different spurious components [3] and this is partially due to the fact that '1 6D '2 . Particular case. Applying the previous expressions to the binary case with '1 D '2 D 0, we have1
.bb/ s1 .t/ D Ae j2 f d t
wT .t/
fd /
(18.59)
From (18.57), the continuous part becomes A2 N .c/ [jWT . f C f d /j2 C jWT . f Ps .bb/ . f / D 4T The discrete part is instead given by A2 N .d/ [jWT . f C f d /j2 C jWT . f Ps .bb/ . f / D 4T 2 C 2Re[WT . f C The behavior of
1 A2 T
f d /j2
2 .h/ Re[WT . f C f d /W . f T
f d /]] (18.60)
f d /j2
C1 X `D 1
(18.61) . f `=T /
f d /W . f T
f d /]]
N .c/ Ps .bb/ . f / is illustrated in Figure 18.24 for three values of the moduN .bb/ lation index; the amplitude in dB of the spectral lines of 12 P .d/ . f / is listed in Table 18.1.
A s
In any case, unless specic assumptions are made on f 1 and f 2 , the phase of the transmitted signal can undergo discontinuities at the instants kT .
1210
Figure 18.24. Continuous part of the power spectral density of a non-coherent binary FSK for three values of the modulation index. Table 18.1 Amplitude in dB of several spectral lines of non-coherent binary FSK signals for three values of the modulation index.
h D 2 fd T fT 0 1 2 3 4 5 6 7 1.0 7:4 15:4 19:2 21:8 23:8 25:4 26:8 1.5 13:4 6:8 20:1 24:6 27:5 29:6 31:3 32:7 2.0 6:0
envelope fsi.bb/ .t/g. As will be derived next, the continuous and discrete parts of the power spectrum are given by 8 2 9 M M = 1 <X .bb/ 1 X .bb/ .c/ N jSi . f /j2 Si . f / Ps .bb/ . f / D (18.62) ; M T : i D1 M i D1
1211
N .d/ Ps .bb/ . f /
1 MT
` T
(18.63)
We derive now the expressions (18.62) and (18.63). s .bb/ is a cyclostationary process with autocorrelation rs .bb/ .t; t / D E[s .bb/ .t/s .bb/ .t D
1 X 1 X
/]
.bb/ k1 T /sak .t
2
.bb/ E[sak .t
1
k2 T /]
k1 D 1 k2 D 1 1 X 1 X
.bb/ E[sak .t
1
.bb/ k1 T /sak .t
2
k2 T /]
(18.64)
.bb/ k1 T /sak .t
1
k1 T /]
Assuming ak1 statistically independent of ak2 for k2 6D k1 , we have that rs .bb/ .t; t / D
1 X 1 X
"
k1 D 1 k D 1 2 k1 6Dk2 1 X k1 D 1
M 1 X .bb/ s .t M i D1 i1
1
M 1 X .bb/ k1 T / s .t M i D1 i2
2
# k2 T /
"
M 1 X .bb/ s .t M i D1 i
k1 T /si.bb/ .t
# k1 T / (18.65)
rs .bb/ .t; t
/ dt
C1 X mD 1
(18.66)
.bb/ .bb/ where r.bb/ is the cross-correlation between s1 and s2 , with Fourier transform s1 s2 .bb/ .bb/ S1 . f /S2 . f /. The second term of (18.66) is the term for m D 0 of the rst summation. Of the three terms in (18.66) we identify the rst as a periodic signal of period T ; it is this term that yields a discrete component in the power spectrum of s .bb/ . Taking the
1212
x. C mT / D repT x. /
C1 1 X ` f ! X T `D 1 T
` T
(18.67)
18.4.2
Coherent demodulator
A coherent demodulator is used when the transmitter provides all M signals with a well dened phase (see Example 6.7.1 on page 486); at the receiver there must be a circuit for the recovery of the phase of the various carriers. In practice this coherent receiver is rarely used because of its implementation complexity. In any case, for orthogonal signals we can adopt the scheme of Figure 6.8, repeated in Figure 18.25 for the binary case. For this case D 0, and the bit error probability has already been derived in (6.71), s ! Es (18.68) Pbit D Q N0 From (6.76), we note that the signals are orthogonal if .2 f d /min D 1 2T (18.69)
1213
Non-coherent demodulator
The transmitted waveforms can now have a different phase, and in any case unknown to the receiver (see Example 6.11.1 on page 509). For orthogonal signals, from the general scheme of Figure 6.59, we give in Figure 18.26 a possible non-coherent demodulator. In this case, for D 0, the bit error probability has already been derived in (6.341), Pbit D 1 e 2
Es 2N0
(18.70)
From (6.306), we note that, in the non-coherent case, to have D 0 it must be .2 f d /min D 1 T (18.71)
Therefore a non-coherent FSK system needs a double frequency deviation and has slightly lower performance (see Figure 6.62) as compared to coherent FSK, however, it does not need to acquire the carrier phase.
Limiter-discriminator FM demodulator
A simple non-coherent receiver, with good performance, for f 1 T 1, f 2 T 1 and sufciently high E s =N0 , is depicted in Figure 18.27. After a passband lter to partially eliminate noise, a classical FM demodulator is used to extract the instantaneous frequency deviation of r.t/, equal to f d . Sampling at instants O t0 C kT and using a threshold detector, we get the detected symbol ak 2 f 1; 1g. In general, for 2 f d T D 1, the performance of this scheme is very similar to that of a non-coherent orthogonal FSK.
1214
18.5
We recall the following characteristics of an ideal FSK signal. In order to avoid high frequency components in the power spectrum of s, the phase of an FSK signal should be continuous and not as represented in Figure 18.23. For signals as given by (18.52), with '1 D '2 , this implies that it must be f n T D In , n D 1; 2, In an integer. To minimize the error probability it must be sn .t/ ? sm .t/, m 6D n. To minimize the required bandwidth, the separation between the various frequencies must be minimum. For signals as given by (18.52), choosing jI1 I2 j D 1 we have f 2 D f 1 C 1=T . In this case s1 and s2 are as shown in Figure 18.28. Moreover, from (18.71) it is easy to prove that the signals are orthogonal. The result is that the frequency deviation is such that 2 f d D j f 2 f 1 j D 1=T , and the carrier frequency is f 0 D . f 1 C f 2 /=2 D f 1 C 1=.2T /.
1
0.8
0.6
0.4
0.2
0
s1(t), s2 (t)
0.2
0.4
0.6
0.8
1
0
T/2 t
Figure 18.28. Waveforms as given by (18.52) with A D 1, f1 D 2=T, f2 D 3=T, and '1 D '2 D 0.
1215
In a digital implementation of an FSK system, the following observation is very useful. Independent of f 1 and f 2 , a method to obtain a phase continuous FSK signal given by s.t/ D A cos.'.t// where '.t/ D 2 f 0 t C ak 2 f d t C k ak 2 f 1; 1g kT t < .k C 1/T (18.73) (18.72)
is to employ a single oscillator, whose phase satises the constraint '..k C 1/T . / / D '..k C 1/T .C/ /; thus it is sufcient that at the beginning of each symbol interval kC1 is set equal to kC1 D ..ak akC1 /2 f d .k C 1/T C k / mod 2 (18.74)
An alternative method is given by the scheme of Figure 18.29, in which the sequence fak g, with binary elements in f 1; 1g, is ltered by g to produce a PAM signal x f .t/ D
C1 X kD 1
ak g.t
kT /
ak 2 f 1; 1g
(18.75)
(18.76)
(18.77)
1216
we obtain a modulation scheme called continuous phase FSK (CPFSK). In turn, CPFSK is a particular case of the continuous phase modulation described in Appendix 18.A. We note that the area of g.t/ is equal to 0.5; in this case the information is in the instantaneous frequency of s, in fact f s .t/ D 1 d'.t/ hak D f 0 C hx f .t/ D f 0 C 2 dt 2T kT < t < .k C 1/T (18.79)
The minimum shift keying (MSK) modulation is equivalent to CPFSK with modulation index h D 0:5, that is for f d D 1=.4T /. Summarizing: an MSK scheme is a binary FSK scheme (Tb D T ) in which, besides having f d D 1=.4T /, the modulated signal has continuous phase. Figure 18.30 illustrates a comparison between an FSK signal with f d D 1=.4T / as given by (18.52), and an MSK signal for a binary sequence fak g equal to f 1; 1; 1; 1g. Note that in the MSK case, it is like having four waveforms s1 ; s1 ; s2 ; s2 , each pair related to ak D 1 and ak D 1, respectively. Signal selection within the pair is done in a way to guarantee the phase continuity of s.
1217
18.5.1
From the expressions given in [3], the behavior of the continuous part of the power spectral density is represented in Figure 18.31. Note that for higher values of h a peak will tend to emerge around f d D h=.2T /, showing the instantaneous frequency of waveforms. Comparing the spectra of Figure 18.31 with those of Figure 18.24 it is seen that phase continuity implies a lower bandwidth, at least if we use the denition of bandwidth based on a given percentage of signal power.
18.5.2
g. / d
(18.80)
is illustrated in Figure 18.32. From (18.76) the phase deviation of the modulated signal is given by 1'.t/ D '.t/ 2 f 0 t D
C1 X kD 1
ak q.t
kTb /
(18.81)
Figure 18.31. Continuous part of the power spectral density of a CPFSK signal for ve values of the modulation index.
1218
q(t) 0.5
Tb
2Tb
3Tb
Let
k 1
k 1 X ai 2 iD 1
(18.82)
t Tb
k C
k 1
(18.83)
Assuming
C k 1D k 2 D 0, the distinct values assumed by k mod 2 are only four: 8 < ; for k even 2 2 k D : 0; for k odd
(18.85)
From (18.84) we note how the information ak is also contained in the variation, in two successive instants, of the phase deviation 1'. The possible trajectories of 1' are shown in Figure 18.33. From (18.81), the complex envelope of the modulated signal is given by s .bb/ .t/ D Ae j1'.t/ (18.86) where 1'.t/ is related to the message by (18.83). Correspondingly the modulated signal (18.76) becomes s.t/ D I .t/ cos.2 f 0 t/ where I .t/ D Re[s .bb/ .t/] The signal s .bb/ .t/ D I .t/ C j Q.t/ (18.89) is represented in Figure 18.34. We note that the phase 1'.t/ continuously varies in time and assumes the values f0; =2; ; 3=2g at instants kTb . Q.t/ D Im[s .bb/ .t/] (18.88) Q.t/ sin.2 f 0 t/ (18.87)
1219
2 3/2 /2 0
(t)
6Tb 0 Tb 2Tb 3Tb 4Tb 5Tb a 0=1 a1 =-1 a 2=1 a 3=1 a 4=1 a5 =-1
(18.90)
wTb .t
(18.91)
1220
from which we can observe that an MSK signal is a binary FSK signal with frequencies f 1 D f 0 1=.4T / and f 2 D f 0 C 1=.4T /, where f 0 is the carrier frequency. In (18.91) the symbols of the sequence fak g, ak 2 f 1; 1g, are information symbols, and k D .ak D .ak In particular we note that 8 > k > > < k k D > k > > : k
1
ak /2
1 kTb C k 4Tb
1
mod 2 (18.92)
ak / k C k 2
mod 2
1 1 1 1
ak D ak ak 6D ak ak D ak ak 6D ak
1 1 1 1
hence k 2 f0; g. From the comparison of (18.90) with (18.83) it is easy to derive the relation between k and k , k D
k 1
kak
(18.94)
MSK as OQPSK
As ak 2 f 1; 1g, it is easy to verify that t t D cos cos ak sin k D 0 2Tb 2Tb Moreover, from (18.91) we obtain I .t/ D A DA X
k
(18.95)
X
k
t kTb / cos ak C k 2Tb t kTb / cos k cos C0 2Tb t kTb / sin ak C k 2Tb t kTb / ak cos k sin C0 2Tb (18.96)
Q.t/ D A DA
X
k
X
k
As the signal s is continuous phase by construction, from (18.87) I and Q must be continuous functions; therefore 1. cos k can change only for k odd, that is at instants .k 2Tb C Tb / in which cos 2Ttb vanishes;
1221
2. ak cos k can change only for k even, that is at instants .k 2Tb / in which sin vanishes.
t 2Tb
Therefore, we note that the information symbols associated with I and Q components can change only every 2Tb and there is a lag Tb between the two branches. Dening the variable ( for k odd cos k ck D (18.97) ak cos k for k even from (18.93) the following relations hold: ck D ak cos k 1 D ak ck 1 cos k 1 for ak D ak ck D cos k 1 for ak 6D ak D ak ak
1 cos k 1
1 1
for k even
(18.98)
D a k ck
for k odd
(18.99)
Indeed, the transformation that maps the bits of fak g into fck g corresponds to a differential encoder given by2 ck D ck
1 ak
ak 2 f 1; 1g
(18.100)
(18.101)
c2`
1 w2Tb .t
(18.102)
Q.t/ D A
C1 X `D 1
c2` w2Tb .t
Another representation of (18.102) is obtained by recognizing the periodic behavior of the waveforms sin. t=2Tb / and cos. t=2Tb / D sin..t C Tb /=2Tb /, illustrated in Figure 18.35. If we window these waveforms to the intervals .0; 2Tb / and . Tb ; Tb /, respectively, in the other intervals it is sufcient to alternate the sign of the encoded symbol, and I .t/ D A
C1 X `D 1
. 1/` c2`
1 w2Tb .t
2`Tb C Tb / sin .t
.t
`2Tb C Tb / 2Tb
(18.103)
Q.t/ D A
C1 X `D 1
2`Tb / sin
2`Tb / 2Tb
1222
0.5
cos( t/2Tb )
0.5
1 T
Tb
2T b t
3T
4T
5T b
0.5
sin( t/2Tb )
0.5
T b
Tb
2T b t
3T b
4T b
5T
Then it follows that an MSK scheme is an OQPSK scheme with modulation interval T D 2Tb and pulse t Tb t (18.104) h T x .t/ D A rect sin 2Tb 2Tb given in Figure 18.36, with Fourier transform HT x . f / D A 4Tb cos.2 f Tb / e .1 .4 f Tb /2 /
j2 f Tb
(18.105)
We note that the transmitted symbols associated with the OQPSK interpretation are encoded with a suitable sign. The example in Table 18.2 shows how the sequence fak g is mapped into fck g and to the data on the I and Q branches. The modulated signals are shown in Figure 18.37; note the phase continuity of s.
j kC1 ck h T x .t
kTb /
(18.106)
1223
ak ck I data Q data
1 1 1
1 1 1
1 1 1
1 1 1
1 1 1
1 1 1
1 1 1
1 1 1
ck D exp
k X j ai 2 iD 1
! D ej
k
(18.107)
using (18.82). Hence an alternative expression for (18.106) is given by s .bb/ .t/ D
C1 X kD 1
ej
h T x .t
kTb /
(18.108)
,
k 1
D ej
jak
(18.109)
1224
1 0.5 0 0.5 1 1
0
I(t)cos(2f t)
I(t)
0.5 0
0.5 1 0 2 t/T 4 6
0.5 1 0 2 t/T 4 6
Figure 18.37. MSK signal for the data sequence of Table 18.2. Using the modulated signals I and Q, s is formed by (18.87).
18.5.3
Interpreting an MSK scheme as a CPFSK scheme, the modulator is as shown in Figure 18.29, with g.t/ D 1 wT .t/ 2Tb b (18.110)
As a whole the modulator is represented in Figure 18.38. Interpreting instead an MSK scheme as an OQPSK scheme, we have the implementation of Figure 18.16 in which T D 2Tb , h T x .t/ is given by (18.104), and data are encoded and changed in sign according to (18.103). Considering its importance and widespread usage, this scheme is illustrated in Figure 18.39.
18.5.4
Among the different demodulation schemes listed in Figure 18.40, we show in Figure 18.41 a differential 1BDD non-coherent demodulator. This scheme is based on the fact that the phase deviation of an MSK signal can vary of =2 between two suitably instants spaced
1225
A cos[(t)] ak Tb
1 w 2Tb Tb
xf (t)
t -
(t)
+ +
s(t)
serial demodulator
differential demodulator
1226
Tb P/S +
decoder
delay
CR /2
STR
^ ak
of Tb (see (18.85)). In any case, the performance for an AWGN channel is that of a DBPSK scheme, but with half the phase variation; hence from (6.163), with E s that becomes E s =2, we get Pbit D 1 e 2
Es 2N0
(18.111)
Note that this is also the performance of a non-coherent orthogonal binary FSK scheme (see (18.70)). In Figure 18.42 we illustrate a coherent (OQPSK type) demodulator that at alternate instants on the I branch and on the Q branch is of the BPSK type. In this case, from
1227
Figure 18.43. Comparison among various error probabilities. Table 18.3 Increment of 0 (in dB) for an MSK scheme with respect to a coherent BPSK demodulator for Pbit D 10 3 .
(6.151), the error probability for decisions on the symbols fck g is given by s ! 2E s Pbit;Ch D Q N0
(18.112)
As it is as if the bits fck g were differentially encoded, to obtain the bit error probability for decisions on the symbols fak g, we use (6.173): Pbit D 4Pbit;Ch
2 3 8Pbit;Ch C 8Pbit;Ch 4 4Pbit;Ch
(18.113)
In Figure 18.43 we show error probability curves for various receiver types. From the graph we note that to obtain an error probability of 10 3 , going from a coherent system to a non-coherent one, it is necessary to increase the value of 0 D E s =N0 as indicated in Table 18.3.
1228
2ak D 1 Q ck D ck Q Q
2.bk bk ak Q
1 /,
D a0 a1 a k Q Q Q Db
1
(18.114)
bk
In other words, the symbol ck is directly related to the information bit bk ; the performance loss due to (18.113) is thus avoided and we obtain Pbit D Pbit;Ch (18.115)
We emphasize that this differential (pre)coding scheme should be avoided if differential non-coherent receivers are employed, because one error in fak g generates a long error O O O sequence in fbk D bk bk 1 g. a Q
18.5.5
From the analogy with OQPSK, the power spectrum of the complex envelope s .bb/ of an MSK signal is directly obtained from (18.105), 8Tb .1 C cos 4 f Tb / 1 N .c/ P .f/D 2 s .bb/ A 2 .1 16 f 2 Tb2 /2 (18.116)
Figure 18.44. Normalized power spectral density of the complex envelope of signals obtained by four modulation schemes.
1229
Modulating both BPSK and QPSK signals with h T x given by a retangular pulse with duration equal to the symbol period, a comparison between the various power spectra is illustrated in Figure 18.44. We note that for limited bandwidth channels, it is convenient to choose h T x of the raised cosine or square root raised cosine type. However, in radio applications the choice of a rectangular pulse may be appropriate, as it generates a signal with a lower peak/average power ratio and therefore is more suitable to be amplied with a power amplier that operates near saturation. Two observations on Figure 18.44 follow. For the same Tb the main lobe of QPSK extends up to 1=T D 0:5=Tb , whereas that of MSK extends up to 1=T D 0:75=Tb ; thus the lobe of MSK is 50% wider than that of QPSK, consequently requiring a larger bandwidth. At high frequencies the spectrum of MSK decays as 1= f 4 , whereas the spectrum of QPSK decays as 1= f 2 .
18.6
18.6.1
GMSK is a variation of MSK in which, to reduce the bandwidth of the modulated signal s, the PAM signal x f is ltered by a Gaussian lter. Consider the scheme illustrated in Figure 18.45, in which we have the following lters: interpolator lter g I .t/ D shaping lter K gG .t/ D p e 2
K 2 t 2 =2
1 wT .t/ 2T
(18.117)
with
2 Bt K Dp ln.2/
FM VCO s(t)
ak
xf (t)
1230
overall lter g.t/ D g I gG .t/ Considering the signals we have: transmitted binary symbols ak 2 f 1; 1g interpolated signal x I .t/ D
C1 X kD 1
(18.119)
(18.120)
ak g I .t
kT / D
C1 X kD 1
ak
1 wT .t 2T
kT /
(18.121)
ak g.t
kT /
(18.122)
t 1
x f . / d
D A cos.2 f 0 t C 1'.t// p
(18.123)
2E s =T .
From the above expressions it is clearRthat the GMSK signal is a frequency modulated t signal with phase deviation 1'.t/ D 1 x f . / d . An important parameter is the 3 dB bandwidth, Bt , of the Gaussian lter. However, a reduction in Bt , useful in making preltering more selective, corresponds to a broadening of the PAM pulse with a consequent increase in the intersymbol interference, as can be noted in the plots of Figure 18.46. Thus a trade-off between the two requirements is necessary. The product Bt T was chosen equal to 0.3 in the GSM and HIPERLAN standards, and equal to 0.5 in the DECT standard (see Appendix 17.A). The case Bt T D 1, i.e. without the Gaussian lter, corresponds to MSK. Analyzing g in the frequency domain we have g.t/ D g I gG .t/ As GI. f / D
1 2
! G. f / D G I . f / G G . f /
(18.124)
sinc. f T /e
2 2 . f =K /2
j f T
(18.125)
GG. f / D e
1231
0.6
0.5
0.4
T g(t)
0.3
B T=
t
Bt T = 0.5
0.2
Bt T = 0.3 Bt T = 0.1
0.1
0 2
1.5
0.5
0.5
1.5
2.5
t/T
Figure 18.46. Overall pulse g.t/ D gI gG .t/, with amplitude normalized to 1=T, for various values of the product Bt T.
it follows that G. f / D
1 2
sinc. f T /e
2 2 . f =K /2
j f T
(18.126)
In Figure 18.47, the behavior of the phase deviation of a GMSK signal with Bt T D 0:3 is compared with the phase deviation of an MSK signal; note that in both cases the phase is continuous, but for GMSK we get a smoother curve, without discontinuities in the slope. Possible trajectories of the phase deviation for Bt T D 0:3 and Bt T D 1 are illustrated in Figure 18.48. Values of e j1'.t/ at the decision instants T C kT are illustrated in Figure 18.49 for a GMSK signal with Bt T D 0:3.
18.6.2
GMSK is a particular case of phase modulation which performs a non-linear transformation on the message, thus it is very difcult to evaluate analytically the power spectrum of a GMSK signal. Hence we resort to a discrete time spectral estimate, for example, the Welch method (see Section 1.11). The estimate is made with reference to the baseband equivalent of a GMSK signal, that is using the complex envelope of the modulated signal given by s .bb/ .t/ D Ae j1'.t/ . The result of the spectral estimate is illustrated in Figure 18.50. Note that the central lobe has an extension up to 0:75=T ; therefore the sampling frequency, FQ , to simulate the baseband equivalent scheme of Figure 18.45 may be chosen equal to FQ D 4=T or FQ D 8=T .
1232
B T = 0.3
t
(t)
8
Bt T =
10
12
14
16
(t)
8
t/T
10
12
14
16
Figure 18.47. Phase deviation 1' of a GMSK signal for Bt T D 0:3, compared with the phase deviation of an MSK signal.
0.8
0.6
0.4
0.2
(t) /
0.2
0.4
0.6
0.8
0.2
0.4
0.6
0.8
1 t/T
1.2
1.4
1.6
1.8
Figure 18.48. Trajectories of the phase deviation of a GMSK signal for Bt T D 0:3 (solid line) and Bt T D 1 (dotted line).
1233
0.8
0.6
0.4
0.2
0.2
0.4
0.6
0.8
1 1
0.8
0.6
0.4
0.2
0 I
0.2
0.4
0.6
0.8
10
B T=0.3
BtT=0.5
10
B T=1.0
t
BtT=
20
PSD (dB)
30
40
50
60
70
80
0.5
1.5
2 fT
2.5
3.5
Figure 18.50. Estimate of the power spectral density of a GMSK signal for various values of Bt T.
1234
18.6.3
For the scheme of Figure 18.45 three possible congurations are given, depending on the position of the DAC.
Conguration I
The rst conguration, illustrated in Figure 18.51, is in the analog domain; in particular, an analog low pass Gaussian lter is employed. As shown in [1], it is possible to implement a good approximation of the Gaussian lter by resorting to simple devices, such as lattice LC lters. However, the weak point of this scheme is represented by the VCO, because in an open loop the voltage/frequency VCO characteristic is non-linear and as a consequence the modulation index can vary even by a factor 10 in the considered frequency range.
Conguration II
The second conguration is represented in Figure 18.52. The digital lter g.nTQ / that approximates the analog lter g.t/ is designed by the window method [4, pag. 444]. For an oversampling factor of Q 0 D 8, letting TQ D T =Q 0 , we consider four lters that are obtained by windowing the pulse g.t/ to intervals .0; T /, . T =2; 3T =2/, . T; 2T /, and . 3T =2; 5T =2/, respectively; the coefcients of the last lter are listed in Table 18.4, using the fact that g.nTQ / has even symmetry with respect to the peak at 4TQ D T =2. A comparison among the frequency responses of the four discrete-time lters and the continuous-time lter is illustrated in Figure 18.53. For a good approximation to the analog lter, the possible choices are limited to the two FIR lters with 23 and 31 coefcients, with support . T; 2T / and . 3T =2; 5T =2/, respectively. From now on we will refer to the lter with 31 coefcients. We note from Figure 18.46 that, for Bt T 0:3, most of the pulse energy is contained within the interval . T; 2T /; therefore the effect of interference does not extend over more than three symbol periods. With reference to Figure 18.52, the lter g is an interpolator lter
ak T
DAC
xI (t)
gG
x f (t)
VCO
s(t)
ak T
xn TQ
DAC
s(t)
VCO
1235
g.nTQ / g.4TQ / g.5TQ / g.6TQ / g.7TQ / g.8TQ / g.9TQ / g.10TQ / g.11TQ / g.12TQ / g.13TQ / g.14TQ / g.15TQ / g.16TQ / g.17TQ / g.18TQ / g.19TQ /
value 0.37119 0.36177 0.33478 0.29381 0.24411 0.19158 0.14168 0.09850 0.06423 0.03921 0.02236 0.01189 0.00589 0.00271 0.00116 0.00046
Figure 18.53. Frequency responses of g.t/ and g.nTQ /, for TQ D T=8, and various lengths of the FIR lters.
1236
from T to T =Q 0 , that can be efciently implemented by using the polyphase representation of the impulse response (see Appendix 1.A). Then, recognizing that x k Q 0 C` D f .`/ .akC2 ; akC1 ; ak ; ak 1 /, the inputoutput relation can be memorized in a RAM or lookup table, and let the input vector [akC2 ; akC1 ; ak ; ak 1 ] address one of the possible 24 D 16 output values; this must be repeated for every phase ` 2 f0; 1; : : : ; Q 0 1g. Therefore, there are Q 0 RAMs, each with 16 memory locations.
Conguration III
The weak point of the previous scheme is again represented by the analog VCO; thus it is convenient to partially implement in the digital domain also the frequency modulation stage. Real-valued scheme. The real-valued scheme is illustrated in Figure 18.54. The samples fu n g are given by u n D u.nTQ / D A cos.2 f 1 nTQ C 1'.nTQ // where f 1 is an intermediate frequency smaller than the sampling rate, f1 D N1 1 N2 TQ (18.128) (18.127)
where N1 and N2 are relatively prime numbers, with N1 < N2 . Thus f 1 TQ D N1 =N2 , and N1 n C 1'n D A cos['n ] (18.129) u n D A cos 2 N2 where 1'n D 1'.nTQ / D Z
nTQ 1 n X iD 1
x. / d '
TQ x.i TQ /
(18.130)
More simply, let xn D x f .nTQ / and X n D X n 1 Cxn ; then it follows that 1'n D TQ X n . Therefore in (18.129) 'n , with carrier frequency f 1 , becomes 'n
. f1 /
D 2
N1 N1 .f / n C TQ X n D 'n 11 C 2 C TQ x n N2 N2
(18.131)
that is the value 'n 1 is obtained by suitably scaling the accumulated values of xn . To .f / obtain u n , we map the value of 'n 1 into the memory address of a RAM which contains values of the cosine function (see Figure 18.55).
.f /
ak T
xn TQ
digital
un TQ
VCO
DAC
u(t)
RF stage
s(t)
1237
Obviously the size of the RAM depends on the accuracy with which u n and 'n are quantized.3 We note that u.nTQ / is a real-valued passband signal, with spectrum centered around the frequency f 1 ; the choice of f 1 is constrained by the bandwidth of the signal u.nTQ /, equal to about 1:5=T , and also by the sampling period, chosen in this example equal to T =8; then it must be 31 C f1 > 0 4T 31 4 C f1 < 4T T (18.132)
or 3=.4T / < f 1 < 13=.4T /. A possible choice is f 1 D 1=.4TQ / assuming N1 D 1 and N2 D 4. With this choice we have a image spacing/signal bandwidth ratio equal to 4/3. Moreover, cos.'n / D cos.2.=4/n C 1'n / becomes cos..=2/n C 1'n /, which in turn is equal to cos.1'n / for n even and sin.1'n / for n odd. Therefore the scheme of Figure 18.54 can be further simplied. Complex-valued scheme. Instead of digitally shifting the signal to an intermediate frequency, it is possible to process it at baseband, thus simplifying the implementation of the DAC equalizer lter. Consider the scheme of Figure 18.56, where at the output of the exponential block we .bb/ .bb/ .bb/ have the sampled signal s .bb/ .nTQ / D Ae j1'.nTQ / D sI .nTQ /C jsQ .nTQ /, where sI
To avoid quantization effects, the number of bits used to represent the accumulated values is usually much larger than the number of bits used to represent 'n . In practice 'n coincides with the most signicant bits of the accumulated values.
1238
.bb/ and sQ are the in phase and quadrature components. Then we obtain .bb/ sI .nTQ / D A cos.1'n /
.bb/
1'n
(18.133)
Once the two components have been interpolated by the two DACs, the signal s.t/ can be reconstructed as
.bb/ s.t/ D sI .t/ cos.2 f 0 t/ .bb/ sQ .t/ sin.2 f 0 t/
(18.134)
With respect to the real-valued scheme, we still have a RAM which stores the values of the cosine function, but two DACs are now required.
18.6.4
According to Laurent [5], if Bt T 0:3, a GMSK signal can be approximated by a QAM signal given by s .bb/ .t/ D
C1 X kD 1
e j 2 6iD
k
1 ai
h T x .t
kTb / (18.135)
C1 X kD 1
kC1
ck h T x .t
kTb /
c k D ck
1 ak
where h T x .t/ is a suitable real-valued pulse that depends on the parameter Bt T and has a support equal to .L C 1/Tb , if L Tb is the support of g.t/. For example, we show in Figure 18.57 the plot of h T x for a GMSK signal with Bt T D 0:3. The linearization of s .bb/ , that leads to interpreting GMSK as a QAM extension of MSK with a different transmit pulse, is very useful for the design of the optimum receiver, which is the same as for QAM systems. Figure 18.58 illustrates the linear approximation of the GMSK model. As for MSK, also for GMSK it is useful to differentially (pre)code the data fak g if a coherent demodulator is employed.
1239
0.9
0.8
0.7
0.6
hTx(t)
0.5
0.4
0.3
0.2
0.1
0 1
2 t/T
j k+1 ak ck c k j k+1 s
(bb)
c k = c k-1 ak
hT x
(t)
Figure 18.58. Linear approximation of a GMSK signal. hTx is a suitable pulse which depends on the value of Bt T. Table 18.5 Values of coefcient c as a function of the modulation system.
where the coefcient c assumes the values given in Table 18.5 for four modulation systems. The plots of Pbit;Ch for the various cases are illustrated in Figure 18.59. As usual, if the data fak g are differentially (pre)coded, Pbit D Pbit;Ch holds, otherwise the relation (18.113) holds. From now on we assume that a differential precoder is employed in the presence of coherent demodulation.
1240
10
10
/4DQPSK
10
2
Pbit,ch
10
10
10
10
6
(dB)
10
12
14
Figure 18.59. Pbit,Ch as a function of 0 for the four modulation systems of Table 18.5.
For the case Bt T D 0:3, and for an ideal AWGN channel, in Figure 18.60 we also give the performance obtained for a receive lter g Rc of the Gaussian type [7, 8], whose impulse response is given in (18.118), where the 3 dB bandwidth is now denoted by Br . Clearly the optimum value of Br T depends on the modulator type and in particular on Bt T . System performance is evaluated using a 4-state VA or a threshold detector (TD). The VA uses an estimated overall system impulse response obtained by the linear approximation of GMSK. The Gaussian receive lter is characterized by Br T D 0:3, chosen for best performance. We observe that the VA gains a fraction of dB as compared to the TD; furthermore, the performance is slightly better than the approximation (18.136). Non-coherent demodulator (1BDD). For a non-coherent receiver, as illustrated in Figure 18.41, without including the receive lter g A , we illustrate in Figure 18.61 the eye diagram at the decision point of a GMSK system, for three values of Bt T . We note that, for decreasing values of Bt T , the system exhibits an increasing level of ISI, and the eye tends to shut; including the receive lter g A with nite bandwidth this phenomenon is further emphasized. Simulation results obtained by considering a receive Gaussian lter g A , and a Gaussian baseband equivalent system, whose bandwidth is optimized for each different modulation, are shown in Figure 18.62. The sensitivity of the 1BDD to the parameter Bt T of GMSK is higher as compared to a coherent demodulator. With decreasing Bt T , we observe a considerable worsening of performance; in fact, to achieve a Pbit D 10 3 the GMSK with Bt T D 0:5 requires a signal-to-noise ratio 0 that is 2.5 dB higher with respect to MSK, whereas GMSK with
1241
10
TD VA
10
Pbit
10
3
10
(dB)
Figure 18.60. Pbit as a function of 0 for a coherently demodulated GMSK (Bt T D 0:3), for an ideal channel and Gaussian receive lter with Br T D 0:3. Two detectors are compared: 1) four-state Viterbi algorithm and 2) threshold detector.
Figure 18.61. Eye diagram at the decision point of the 1BDD for a GMSK system, for an ideal channel and without the lter gA : (a) Bt T D C1, (b) Bt T D 0:5, (c) Bt T D 0:3.
1242
Figure 18.62. Pbit as a function of 0 obtained with the 1BDD for GMSK, for an ideal channel and Gaussian receive lter having a normalized bandwidth Br T.
Bt T D 0:3 requires an increment of 7.8 dB. In Figure 18.62 we also show for comparison purposes the performance of =4-DQPSK with a receive Gaussian lter. The performance of another widely used demodulator, LDI (see Section 18.2.3), is quite similar to that of 1BDD, showing a substantial equivalence between the two non-coherent demodulation techniques applied to GMSK and =4-DQPSK. Comparison. Always for an ideal AWGN channel, a comparison among the various modulators and demodulators is given in Table 18.6. As a rst observation, we note that a coherent receiver with an optimized Gaussian receive lter provides the same performance,
Table 18.6 Required values of 0, in dB, to achieve a Pbit D 10 3 for various modulation and demodulation schemes.
Modulation
=4-DQPSK o QPSK ( D 0:3) MSK GMSK (Bt T D 0:5) GMSK (Bt T D 0:3)
1243
evaluated by the approximate relation (18.136), of the MAP criterion. Furthermore, we note that for GMSK with Bt T 0:5, because of strong ISI, a differential receiver undergoes a substantial penalty in terms of 0 to achieve a given Pbit with respect to a coherent receiver; this effect is mitigated by canceling ISI by suitable equalizers [9]. Another method is to detect the signal in the presence of ISI, in part due to the channel and in part to the differential receiver, by the Viterbi algorithm. Substantial improvements with respect to the simple threshold detector are obtained, as shown in [10]. In the previous comparison between amplitude modulation (=4-DQPSK) and phase modulation (GMSK) schemes we did not take into account the non-linearity introduced by the power amplier (see Section 4.8), which leads to: 1) signal distortion and 2) spectral spreading that creates interference in adjacent channels. Usually, the latter effect is dominant and is controlled using a HPA with a back-off that can be even of several dB. In some cases signal predistortion before the HPA allows a decrease of the OBO. Overall, the best system is the one that achieves, for the same Pbit , the smaller value of .0/d B C .OBO/d B (18.137)
In other words (18.137), for the same Pbit , additive channel noise and transmit HPA, selects the system for which the transmitted signal has the lowest power. Obviously in (18.137) 0 depends on the OBO; at high frequencies, where the HPA usually introduces large levels of distortion, the OBO for a linear modulation scheme may be so large that a phase modulation scheme may be the best solution in terms of (18.137).
Power delay prole HT 0.02 0.75 0.08 0.02 0.01 0.12 UA 0.00566 0.01725 0.80256 0.15264 0.01668 0.00521 RA 0.00086 0.99762 0.00109 0.00026 0.00011 0.00006
1244
100
90
80
70
VA(32) MF+DFE(13,7)
P cdf
bit
60
50
VA(32)
40
30
20
10
3
MF+DFE(13,7)
10
bit
Figure 18.63. Comparison between Viterbi algorithm and DFE preceded by a MF for the multipath channel EQ6, in terms of BER cdf.
The channel model is also obtained by considering a T -spaced impulse response; in particular, the performance is evaluated for the four models given in Table 18.7: equal gain (EQ), hilly terrain (HT), urban area (UA), rural area (RA). The difference in performance between the two receivers is higher for the EQ channel; this is the case shown in Figure 18.63 for two values of the signal-to-noise ratio 0 at the receiver.
Bibliography
[1] B. Razavi, RF microelectronics. Englewood Cliffs, NJ: Prentice-Hall, 1997. [2] K. Feher, Wireless digital communications. Upper Saddle River, NJ: Prentice-Hall, 1995. [3] S. Benedetto and E. Biglieri, Principles of digital transmission with wireless applications. New York: Kluwer Academic Publishers, 1999. [4] A. V. Oppenheim and R. W. Schafer, Discrete-time signal processing. Englewood Cliffs, NJ: Prentice-Hall, 1989. [5] P. Laurent, Exact and approximate construction of digital phase modulations by superposition of amplitude modulated pulses (AMP), IEEE Trans. on Communications, vol. 34, pp. 150160, Feb. 1986.
18. Bibliography
1245
[6] S. Ohmori, H. Wakana, and S. Kawase, Digital communications technologies. Boston, MA: Artech House, 1998. [7] K. Murota and K. H. Hirade, GMSK modulation for digital mobile radio telephony, IEEE Trans. on Communications, vol. 29, p. 1045, July 1991. [8] M. K. Simon and C. C. Wang, Differential detection of Gaussian MSK in a mobile radio environment, IEEE Trans. on Vehicular Technology, vol. 33, pp. 311312, Nov. 1984. [9] N. Benvenuto, P. Bisaglia, A. Salloum, and L. Tomba, Worst case equalizer for noncoherent HIPERLAN receivers, IEEE Trans. on Communications, vol. 48, pp. 2836, Jan. 2000. [10] N. Benvenuto, P. Bisaglia, and A. E. Jones, Complex noncoherent receivers for GMSK signals, IEEE Journal on Selected Areas in Communications, vol. 17, pp. 18761885, Nov. 1999.
1246
Appendix 18.A
A signal with constant envelope can be dened by its passband version as r 2E s s.t/ D cos.2 f 0 t C 1'.t; a// T
(18.138)
where E s is the energy per symbol, T the symbol period, f 0 the carrier frequency and a denotes the symbol message fak g at the modulator input. For a continuous phase modulation scheme, the phase deviation 1'.t; a/ can be expressed as Z t 1'.t; a/ D 2 h x f . / d (18.139)
1
with x f .t/ D
C1 X kD 1
ak g.t
kT /
(18.140)
where g.t/ is called instantaneous frequency pulse. In general, the pulse g.t/ satises the following properties: limited duration symmetry normalization
0
g. / d D
ai g.
i T / d D 2 h
k X iD 1
Z ai
t iT 1
g. / d
1 iD 1
(18.144)
q.t
iT/
kT t < .k C 1/T
(18.145)
g. / d
(18.146)
1247
1/2
1/2
1/2
The pulse q.t/ is called phase response pulse and represents the most important part of the CPM signal because it indicates to what extent each information symbol contributes to the overall phase deviation. In Figure 18.64 the phase response pulse q.t/ is plotted for PSK, CPFSK, and BFSK. In general the maximum value of the slope of the pulse q.t/ is related to the width of the main lobe of the PSD of the modulated signal s.t/, and the number of continuous derivatives of q.t/ inuences the shape of the secondary lobes. In general, information symbols fak g belong to an M-ary alphabet A, that for M even is given by f1; 3; : : : ; .M 1/g. The constant h is called modulation index and determines, together with the dimension of the alphabet, the maximum phase variation in a symbol period, equal to .M 1/h . By changing q.t/ (or g.t/), h, and M, we can generate several continuous phase modulation schemes. The modulation index h is always given by the ratio of two integers, h D `= p, because this implies that the phase deviation, evaluated modulo 2 , assumes values in a nite alphabet. In fact, we can write 1'.t; a/ D 2 h with
k L k X i Dk LC1
ai q.t
iT/ C
k L
kT t < .k C 1/T
(18.147)
X ` k L ai D p iD 1
"
# (18.148)
mod 2
k L is called phase state; it represents the overall contribution given by the symbols : : : ; a 1 ; a0 ; a1 ; : : : ; ak L to the phase duration in the interval [kT; .k C 1/T /, and can only assume 2 p distinct values. The rst term in (18.147) is called corrective state and, because it depends on L symbols ak LC1 ; : : : ; ak , at a certain instant t it can only assume M L 1 distinct values. The phase deviation is therefore characterized by a total number of values equal to 2 pM L 1 . CPM schemes with L D 1, i.e. CPFSK, are called full response schemes and have a reduced complexity with a number of states equal to 2 p; schemes with L 6D 1 are instead
1248
called partial response schemes. Because of the memory in the modulation process a partial response scheme allows a trade-off between the error probability at the receiver and the shaping of power spectra of the modulated signal. However, this advantage is obtained at the expense of a greater complexity of the receiver. For this reason the modulation index is usually a simple rational number as 1; 1=2; 1=4; 1=8.
Advantages of CPM
The popularity of the continuous phase modulation technique derives rst of all from the constant envelope property of the CPM signal; in fact, a signal with a constant envelope allows using very efcient power ampliers. In the case of linear modulation techniques, as QAM or OFDM, it is necessary to compensate for the non-linearity of the amplier by predistortion or to decrease the average power in order to work in linear conditions. Before the introduction of TCM, it was believed that source and/or channel coding would allow an improvement in performance only at the expense of a loss in transmission efciency, and hence would require a larger bandwidth; CPM permits both good performance and highly efcient transmission. However, one of the drawbacks of the CPM is the implementation complexity of the optimum receiver, especially in the presence of dispersive channels.
Chapter 19
Design of high speed transmission systems over unshielded twisted pair cables
In this chapter we describe the design of two high speed data transmission systems over unshielded twisted pair cables [1, 2].
19.1
Design of a quaternary partial response class-IV system for data transmission at 125 Mbit/s
Figure 19.1 shows the block diagram of a transceiver for a quaternary partial response class-IV (QPR-IV), or quaternary modied duobinary (see Appendix 7.A), system for data transmission at 125 Mbit/s over unshielded twisted-pair cables [1]. In the transmitter, information bits are rst scrambled and then input to a 2B1Q differential encoder that yields output symbols belonging to the quaternary alphabet A D f 3; 1; C1; C3g (see Example 6.5.1 on page 479); differential encoding makes the transmission of information insensitive to the polarity of the received signals. Signal shaping into PR-IV form is accomplished by the cascade of the following elements: the D/A converter, the analog transmit lter (ATF), the cable, the analog receive lter (ARF) with automatic gain control (AGC), the A/D converter, and, in the digital domain, the xed decorrelation lter (DCF) and the adaptive equalizer. After equalization, the sequence of transmitted quaternary symbols is detected by a Viterbi algorithm. As the Viterbi algorithm introduces a delay in the detection of the transmitted sequence, as an alternative it is possible to use a threshold detector that individually detects transmitted symbols with a negligible delay. If the mean-square error (MSE) is below a xed value, threshold detection, or symbol-by-symbol detection (SSD), is selected as the corresponding symbol decisions are then sufciently reliable. If, instead, the MSE exceeds the xed value, the Viterbi algorithm is employed. Finally, differential 1Q2B decoding and descrambling are performed.
1250
Figure 19.1. Block diagram of a QPR-IV transceiver. [From [1], c 1995 IEEE.]
Figure 19.2. Overall analog channel considered for the joint optimization of the analog transmit (ATF) and receive (ARF) lters. [From [1], c 1995 IEEE.]
complexity, the transfer functions of these lters are rst expressed in terms of poles and zeros; the pole-zero congurations are then jointly optimized by simulated annealing as described in [3]. The cost function that is used for the optimization reects two criteria: a) the mean-square error between the impulse response h.t; L ; u c / for a cable length equal to 50 m, and the ideal PR-IV response must be below a certain value and b) spectral components of the transmitted signal above 30 MHz should be well suppressed to achieve compliance with regulations on radiation limits. Investigations with various orders of the respective transfer functions have shown that a good approximation of an ideal PR-IV response is obtained with 5 poles and 3 zeros for the ATF, and 3 poles for the ARF.
ak h[t
kT; L ; u c .L/] C
C1 X kD 1
N ak h N [t
(19.1)
1251
N where fak g and fak g are the sequences of quaternary symbols generated by the local transmitter and remote transmitter, respectively, h N [t; L ; u c .L/] is the NEXT channel response and w R .t/ is additive Gaussian noise. The signal x.t/ is sampled by the ADC that operates synchronously with the DAC at the modulation rate of 1=T D 62:5 MBaud. The adjustment of the AGC circuit is computed digitally, using the sampled signal x k D x.kT /, such that the ADC output signal achieves a constant average statistical power M R , i.e., Z 1 T 2 2 D E[x .t/] dt D MR (19.2) E[x k ] u c Du c .L/ T 0 u c Du c .L/
This ensures that the received signal is converted with optimal precision independently of the cable length; moreover, a controlled level of the signal at the adaptive digital equalizer input is required for achieving optimal convergence properties.
(19.3)
0 i NN
(19.4)
where N is the adaptation gain. As discussed in Chapter 16, high-speed full duplex transmission over two separate wire pairs with NEXT cancellation and full-duplex transmission over a single pair with echo cancellation pose similar challenges. In the latter case, a hybrid is included to separate the two directions of transmission; a QPR-IV transceiver with NEXT cancellation can then be used also for full-duplex transmission over a single pair, as in this case the NEXT canceller acts as an echo canceller.
Decorrelation lter
After NEXT cancellation, the signal is ltered by a decorrelation lter, which is used to improve the convergence properties of the adaptive digital equalizer by reducing the correlation between the samples of the sequence fx k g. The ltering operation performed by the Q DCF represents an approximate inversion of the PR-IV frequency response. The DCF has frequency response 1=.1 e j4 f T /, with 0 < < 1, and provides at its output the signal Q z k D xk C z k
2
(19.5)
1252
Adaptive equalizer
The samples fz k g are stored in an elastic buffer, from which they are transferred into the equalizer delay line. Before describing this operation in more detail, we make some observations about the adaptive equalizer. As mentioned above, the received signal x.t/ is sampled in synchronism with the timing of the local transmitter. Due to the frequency offset between local and remote transmitter clocks, the phase of the remote transmitter clock will drift in time relative to the sampling phase. As the received signal is bandlimited to one half of the modulation rate, signal samples taken at the symbol rate are not affected by aliasing; hence, a fractionally-spaced equalizer is not required for a QPR-IV system. Furthermore, as the signal value x.t/ can be reconstructed from the T -spaced samples fz k g, an equalizer of sufcient length acts also as an interpolator. The adaptive equalizer output signal is given by yk D
E where fci;k g, i D 0; : : : ; N E NE 1 X i D0 E ci;k z k i
(19.6)
1253
E zk
O i ek
0 i NE
(19.7)
O O O where ek D yk .ak ak 2 / is the error obtained using tentative decisions ak on the O transmitted quaternary symbols, and E is the adaptation gain.
Figure 19.3. Convergence of the adaptive NEXT canceller. [From [1], c 1995 IEEE.]
1254
Figure 19.4. Convergence of the adaptive equalizer for (a) best-case sampling phase and (b) worst-case sampling phase. [From [1], c 1995 IEEE.]
1255
Figure 19.5. Convergence of the adaptive equalizer for worst-case timing phase drift. [From [1], c 1995 IEEE.]
is also indicated. For the simulations, the NEXT canceller was assumed to be realized in the distributed-arithmetic form (see Section 16.1). The convergence of the MSE at the output of the equalizer in the absence of timing phase drift is shown in Figure 19.4a and b for best and worst-case sampling phase, respectively, and a value 0T x D 2E T x =N0 of 43 dB, where E T x is the average energy per modulation interval of the transmitted signal. The NEXT canceller is assumed to have converged to the optimum setting. An equalizer length of N E D 24 is chosen, which guarantees that the mean-square interpolation error with respect to an ideal QPR-IV signal is less than 25 dB for the worst-case sampling phase. The self-training period is TST 400 s, corresponding to approximately 25000T . The adaptation gains for self-training and decision directed adjustment have the same value E D 2 9 , that is chosen for best performance in the presence of a worst-case timing phase drift D 10 4 . Figure 19.5 shows the meansquare error convergence curves obtained for D 10 4 .
19.1.1
VLSI implementation
1256
from the received signal, look-up values are selected by the bits in the NEXT canceller delay line and added by a carry-save adder. By segmenting the delay line of the NEXT canceller into sections of shorter lengths, a trade-off concerning the number of operations per modulation interval and the number of memory locations that are needed to store the look-up values is possible. The convergence of the look-up values to the optimum setting is achieved by an LMS algorithm. If the delay line of the NEXT canceller is segmented into L sections with K D N N =L delay elements each, the NEXT canceller output signal is given by uk D ON
NN 1 X i D0 N N ak i ci;k D L 1 X X K 1 `D0 mD0 N ak N `K m c`K Cm;k
(19.8)
N In a distributed-arithmetic implementation, the quaternary symbol ak is represented by .0/ .1/ the binary vector [ak ; ak ], that is N ak D 1 X wD0 .w/ .2ak
1/2w D
1 X wD0
.w/ bk 2w
(19.9)
.w/ .w/ 1/ 2 f 1; C1g. Introducing (19.9) into (19.8) we obtain (see (16.12)) where bk D .2ak " # L 1 1 K 1 X X X .w/ N uk D ON 2w bk `K m c`K Cm;k (19.10) `D0 wD0 mD0
Equation (19.10) suggests that the lter output can be computed using a set of L2 K look-up values that are stored in L look-up tables with 2 K memory locations each. Extracting the .w/ term bk `K out of the square bracket in (19.10), to determine the output of a distributedarithmetic lter with reduced memory size L2 K 1 we rewrite (19.10) as (see (16.13)) uk D ON
L 1 1 X X `D0 wD0 .w/ N .w/ 2w bk `K dk .i k;` ; `/
(19.11)
.w/ N where fdk .n; `/g, n D 0; : : : ; 2 K 1 1, ` D 0; : : : ; L 1, are the look-up values, and i k;` denotes the selected look-up address that is computed as follows: 8 > K 1 .w/ > X .w/ > a 2m 1 if ak `K D 1 > > < mD1 k `K m .w/ i k;` D (19.12) >K 1 > X .w/ > .w/ > ak `K m 2m 1 N if ak `K D 0 > : mD1
where an is the ones complement of an . The expression of the LMS algorithm to update N the look-up values of a distributed-arithmetic NEXT canceller takes the form
N N dkC1 .n; `/ D dk .n; `/ C N x k Q 1 X wD0 .w/ 2w bk `K n
.w/ i k;`
(19.13) 1
n D 0; : : : ; 2 K
` D 0; : : : ; L
1257
where n is the Kronecker delta. We note that at each iteration only those look-up values that are selected to generate the lter output are updated. The implementation of the NEXT canceller is further simplied by updating at each iteration only the look-up values that are addressed by the most signicant bits of the symbols, i.e. those with index w D 1, stored in the delay line (see (16.23)). The block diagram of an adaptive distributed-arithmetic NEXT canceller is shown in Figure 19.6. In the QPR-IV transceiver, for the implementation of a NEXT canceller with a time span of 48T , L D 16 segments with K D 3 delay elements each are employed. The look-up values are stored in 16 tables with four 16-bit registers each.
ak(0)
ak(1)
address computation
(1) i k,0
a k(L1)K
(0)
a k(L1)K
(1)
address computation
(1) i k,L1
ik,0
(0)
i k,L1
(0)
~k x
N (1) d k+1 (i k,0 ,0)
table 0
~ xk
table L1 +
1 0
+1 1
bk(1)
1 0
+1 1
b k(L1)K
(1)
1 0
+1 1
bk(0)
1 0
~ x
^ ukN
+
Figure 19.6. Adaptive distributed-arithmetic NEXT canceller. [From [1], c 1995 IEEE.]
+1 1
b k(L1)K
(0)
xk
1258
ciE z k
.i 1/
(19.14)
This MAC unit is then reset and its output will be considered again N E time instants later. Figure 19.7 depicts the implementation of the digital equalizer. The N E coefcients fciE g, i D 0; : : : ; N E 1, normally circulate in the delay line shown at the top of the gure. Except when recentering of the equalizer coefcients is needed, N E coefcients in the delay line are presented each to a different MAC unit, and the signal sample z k is input to all MAC units. At the next time instant, the coefcients are cyclically shifted by one position and the new signal sample z kC1 is input to all the units. The multiplexer shown at the bottom of the gure selects in turn the MAC unit that provides the equalizer output signal. To explain the operations for recentering of the equalizer coefcients, we consider as an example a simple equalizer with N E D 4 coefcients; in Figure 19.8, where for simplicity the coefcients are denoted by fc0 ; c1 ; c2 ; c3 g, the coefcients and signal samples at the input of the 4 MAC units are given as a function of the time instant. At time instant k, the output of the MAC unit 0 is selected, at time instant k C 1 the output of the MAC unit 1 is selected, and so on. For a negative frequency offset between the local and remote transmitter clocks, a recentering operation corresponding to a left shift of the equalizer coefcients occasionally occurs as illustrated in the upper part of Figure 19.8. We note that as a result of this operation, a new coefcient c4 , initially set equal to zero, is introduced. We also note that signal samples with proper delay need to be input to the MAC units. A similar operation occurs for a right shift of the equalizer coefcients, as illustrated in the lower part of the gure; in this case a new coefcient c 1 , initially set equal to zero, is introduced. In the equalizer implementation shown in Figure 19.7, the control operations to select the lter coefcients and the signal samples are implemented by the multiplexer
1259
10 11 01 00
MUXC Nu coefficient updating terms
E cN
E 1
E c0
c1E + +
cE A
NE 3
cE C
NE 2
z k1 zk
1 0
MUXS0 MUXS1
10
MUXS2
1 0
MUXS N
10
E 2
10
MUXS NE 1
(0) zk
zk
(1)
MAC 0
MAC 1
MAC 2
MAC NE 2
MAC NE 1
MUX
Figure 19.7. Digital adaptive equalizer: coefcient circulation and updating, and computation of output signal. [From [1], c 1995 IEEE.]
MUXC at the input of the delay line and by the multiplexers MUXS(0), : : : , MUXS(N E 1) at the input of the MAC units, respectively. A left or right shift of the equalizer coefcients is completed in N E cycles. To perform a left shift, in the rst cycle the multiplexer MUXC E is controlled so that a new coefcient c N E D 0 is inserted into the delay line. During the following .N E 1/ cycles, the input of the delay line is connected to point B. After E inserting the coefcient c1 at the N E -th cycle, the input of the delay line is connected to point C and normal equalizer operations are restored. For a right shift, the multiplexer MUXC is controlled so that during the rst N E 1 cycles the input of the delay line is connected to point A. A new coefcient c E 1 D 0 is inserted into the delay line at the N E -th cycle and normal equalizer operations are thereafter restored. At the beginning of the equalizer operations, the equalizer coefcients are initialized by inserting the sequence f0; : : : ; 0; C1; 0; 1; 0; : : : ; 0g into the delay line. The adaptation of the equalizer coefcients in decision-directed mode is performed according to the LMS algorithm (19.7). However, to reduce implementation complexity, equalizer coefcients are not updated at every cycle; during normal equalizer operations, each coefcient is updated every N E =NU cycles by adding correction terms at NU equally spaced xed positions in the delay line, as shown in Figure 19.9. The architecture adopted for the computation of the correction terms is similar to the architecture for the computation
1260
MAC 0
c3 c2 c1 c0 c3 c2 c1 z k3 z k2 z k1 zk z k+1 z k+2 z k+3 z k+4 z k+4 z k+5 z
k+6
MAC 1
c0 c3 c2 c1 c0 c3 c2 c1 c0 c4 c3 c2 c1 c4 c3 c2 c1 z k3 z k2 z k1 zk z k+1 z k+2 z k+3 z k+4 z k+5 z z k+6 z k+7 z k+8 z k+9 z k+10 z k+11 z k+12
MAC 2
c1 c0 c3 c2 c1 c0 c3 c2 c1 c0 c4 c3 c2 c1 c4 c3 c2 c1 z k3 z k2 z k1 zk z k+1 z k+2 z k+3 z k+4 z k+5 z
k+6
MAC 3
c2 c1 c0 c3 c2 c1 c0 c3 c2 c1 c0 c4 c3 c2 c1 c4 c3 c2 c1 z k3 z k2 z k1 zk z k+1 z k+2 z k+3 z k+4 z k+5 z
k+6
k3 k2 k1 k k+1 k+2 k+3 k+4 k+5 k+6 k+7 k+8 k+9 k+10 k+11 k+12 k+13 k+14 k+15
shift left
c0 c4 c3 c2 c1 c4 c3 c2 c1
k+5
c3 c2 c1 c0 c3 c2 c1
z k3 z k2 z k1 zk z k+1 z k+2 z k+3 z k+4 z k+6 z k+7 z k+8 z k+9 z k+10 z k+11
c0 c3 c2 c1 c0 c3 c2 c1 c0 c2
c1 c0 c3 c2 c1 c0 c3 c2 c1 c0 c2 c1 c0 c1 c2
z k3 z k2 z k1 zk z k+1 z k+2 z k+3 z k+4 z k+5 z k+6 z k+8 z k+9 z k+10 z k+11
c2 c1 c0 c3 c2 c1 c0 c3 c2 c1 c0 c2
k3 k2 k1 k k+1 k+2 k+3 k+4 k+5 k+6 k+7 k+8 k+9 k+10 k+11 k+12 k+13 k+14 k+15
shift right
c0 c2 c1 c0 c1 c2
c1 c0 z k+12 c1 z k+13
Figure 19.8. Coefcients and signals at the input of the multiply-accumulate (MAC) units during coefcient shifting. [From [1], c 1995 IEEE.]
of the equalizer output signal. The gradient components E z k i ek , i D 0; : : : ; N E 1, O are accumulated in N E MAC units, as illustrated in the gure. The delay line stores the .0/ .1/ signal samples z k and z k . The inputs to each MAC unit are given by the error signal and a signal from the delay line. The multiplexers at the input of each register in the delay line allow selecting the appropriate inputs to the MAC units in connection with the recentering of the equalizer coefcients. At each cycle, the output of NU MAC units are selected and
1261
z (0)
k
10 00 01
z (1)
k
10 00 01
10 00 01
10 00 01
10 00 01
^ ek
MAC 0
MAC 1
MAC 2
MAC NE 2
MAC
NE 1
MUX
Figure 19.9. Adaptive digital equalizer: computation of coefcient adaptation. [From [1], c 1995 IEEE.]
input to the NU adders in the delay line where the equalizer coefcients are circulating, as illustrated in Figure 19.7.
Timing control
An elastic buffer is provided at the boundary between the transceiver sections that operate at the transmit and receive timings. The signal samples at the output of the DCF are obtained at a rate that is given by the transmit timing and stored at the same rate into the elastic buffer. Signal samples from the elastic buffer are read at the same rate that is given by the receive timing. The VCO that generates the receive timing signal is controlled in order to prevent buffer underow or overow. Let WPk and RPk denote the values of the two pointers that specify the write and read addresses, respectively, for the elastic buffer at the k-th cycle of the receiver clock. We consider a buffer with eight memory locations, so that WPk ; RPk 2 f0; 1; 2; : : : ; 7g. The write pointer is incremented by one unit at every cycle of the transmitter clock, while the read pointer is also incremented by one unit at every cycle of the receiver clock. The difference pointer, DPk D WPk RPk .mod 8/ (19.15)
1262
is used to generate a binary control signal 1k of the VCO must be increased or decreased: 8 > C1 > < 1 1k D > > : 1k 1
2 f1g that indicates whether the frequency if DPk D 4; 5 if DPk D 2; 3 otherwise (19.16)
The signal 1k is input to a digital loop lter which provides the control signal to adjust the VCO. If the loop lter comprises both a proportional and an integral term, with corresponding gains of and 1 , respectively, the resulting second-order phase-locked loop is described by (see Section 14.7) kC1 D k C 1k C 1k 1kC1 D 1k C 1 1k (19.17)
where k denotes the difference between the phases of the transmit and receive timing signals. With a proper setting of the gains and 1 , the algorithm (19.17) allows for correct initial frequency acquisition of the VCO and guarantees that the write and read pointers do not overrun each other during steady-state operations. For every time instant k, the two consecutive signal samples stored in the memory locations with the address values RPk and .RPk 1/ are read and transferred to the equalizer. These signal samples are denoted by z k and z k 1 in Figure 19.10. When a recentering of the equalizer coefcients has to take place, for one cycle of the receiver clock the read pointer is either not incremented (left shift), or incremented by two units (right shift). These operations are illustrated in the gure, where the elastic buffer is represented as a circular memory. We note that by the combined effect of the timing control scheme and the recentering of the adaptive equalizer coefcients the frequency of the receive timing signal equals on average the modulation rate at the remote transceiver.
Viterbi detector
For the efcient implementation of near MLSD of QPR-IV signals, we consider the reducedstate Viterbi detector of Example 8.12.1 on page 687. In other words, the signal samples at the output the of .1 D 2 / partial response channel are viewed as being generated by two interlaced .1 D 0 / dicode channels, where D 0 D D 2 corresponds to a delay of two modulation intervals, 2T . The received signal samples are hence deinterlaced into even and odd time-indexed sequences. The Viterbi algorithm using a 2-state trellis is performed independently for each sequence. This reduced-state Viterbi algorithm retains at any time instant k only the two states with the smallest and second smallest metrics and their survivor sequences, and propagates the difference between these metrics instead of two metrics. Because the minimum distance error events in the partialresponse trellis lead to quasi-catastrophic error propagation, a sufciently long path memory depth is needed. A path memory depth of 64T has been found to be appropriate for this application.
1263
Normal operation
WPk
Shift left
WPk
WPk+1 WP k+2
RP k zk Shift right
z k1 RPk+1 zk
z k1
RPk+2 z k+1 zk
WPk
RPk+2 WP k+2
z k+2
19.2
We now describe the 100BASE-T2 system for fast Ethernet mentioned in Section 17.1.2 [4, 2].
1264
Figure 19.11. Dual duplex transmission over two wire pairs. [From [2], c 1997 IEEE.]
from other transmissions in multi-pair cables and far-end cross-talk (FEXT), although normally not very signicant, requires specic structures (see, for example, Section 16.4). To achieve best performance for data transmission over UTP-3 cables, signal bandwidth must be conned to frequencies not exceeding 30 MHz. As shown in Chapter 17, this restriction is further mandated by the requirement to meet FCC and CENELEC class B limits on emitted radiation from communication systems. These limits are dened for frequencies above 30 MHz. Twisted pairs used in UTP-3 cables have fewer twists per unit of length and generally exhibit a lower degree of homogeneity than pairs in UTP-5 cables; therefore transmission over UTP-3 cables produces a higher level of radiation than over UTP-5 cables. Thus it is very difcult to comply with the class B limits if signals containing spectral components above 30 MHz are transmitted over UTP-3 cables. As illustrated in Figure 19.11, for 100BASE-T2 a dual duplex baseband transmission concept was adopted. Bidirectional 100 Mbit/s transmission over two pairs is accomplished by full duplex transmission of 50 Mbit/s streams over each of two wire pairs. The lower modulation rate and/or spectral modulation efciency required per pair for achieving the 100 Mbit/s aggregate rate represents an obvious advantage over mono duplex transmission, where one pair would be used to transmit only in one direction and the other to transmit only in the reverse direction. Dual duplex transmission requires two transmitters and two receivers at each end of a link, as well as separation of the simultaneously transmitted and received signals on each wire pair. Sufcient separation cannot be accomplished by analog hybrid circuits only. In 100BASE-T2 transceivers it is necessary to suppress residual echoes returning from the hybrids and impedance discontinuities in the cable as well as self NEXT by adaptive digital echo and NEXT cancellation. Furthermore, by sending transmit signals with nearly 100% excess bandwidth, received 100BASE-T2 signals exhibit spectral redundancy that can be exploited to mitigate the effect of alien NEXT by adaptive digital equalization. It will be shown later in this chapter that, for digital NEXT cancellation and equalization as well as echo cancellation in the case of dual-duplex transmission, dual-duplex and mono-duplex schemes require a comparable number of multiply-add operations per second.
1265
The dual transmitters and receivers of a 100BASE-T2 transceiver will henceforth be referred to simply as transmitter and receiver. Signal transmission in 100BASE-T2 systems takes place in an uninterrupted fashion over both wire pairs in order to maintain timing synchronization and the settings of adaptive lters at all times. Quinary pulse-amplitude baseband modulation at the rate of 25 MBaud is employed for transmission over each wire pair. The transmitted quinary symbols are randomized by side-stream scrambling. The redundancy of the quinary symbol sets is needed to encode 4-bit data nibbles, to send between data packets an idle sequence that also conveys information about the status of the local receiver, and to insert special delimiters marking the beginning and end of data packets.
Auto-Negotiation Process
TIMEOUT 100BASE-T2 available on both sides of the link
TRAINING STATE: blind receiver training, followed by decision-directed training; send idle
loc_rcvr_status=NOT_OK
loc_rcvr_status=OK rem_rcvr_status=OK
loc_rcvr_status=OK rem_rcvr_status=NOT_OK
Figure 19.12. State diagram of 100BASE-T2 physical layer control. [From [2], c 1997 IEEE.]
1266
received from the remote transmitter would generally shift in phase relative to the alsoreceived echo and self-NEXT signals, as discussed in the previous section. To cope with this effect some form of interpolation would be required, which can signicantly increase the transceiver complexity. After auto-negotiation is completed, both 100BASE-T2 transceivers enter the TRAINING state. In this state a transceiver expects to receive an idle sequence and also sends an idle sequence, which indicates that its local receiver is not yet trained (loc_rcvr_status =NOT_OK). When proper local receiver operation has been achieved by blind training and then by further decision-directed training, a transition to the IDLE state occurs. In the IDLE state a transceiver sends an idle sequence expressing normal operation at its receiver (loc_rcvr_status = OK) and waits until the received idle sequence indicates correct operation of the remote receiver (rem_rcvr_status = OK). At this time a transceiver enters the NORMAL state, during which data nibbles or idle sequences are sent and received as demanded by the higher protocol layers. The remaining transitions shown in the state diagram of Figure 19.12 mainly dene recovery functions. The medium independent interface (MII) between the 100BASE-T2 physical layer and higher protocol layers is the same as for the other 10/100 Mbit/s IEEE 802.3 physical layers. If the control line TX_EN is inactive, the transceiver sends an idle sequence. If TX_EN is asserted, 4-bit data nibbles TXD(3:0) are transferred from the MII to the transmitter at the transmit clock rate of 25 MHz. Similarly, reception of data results in transferring 4-bit data nibbles RXD(3:0) from the receiver to the MII at the receive clock of 25 MHz. Control line RX_DV is asserted to indicate valid data reception. Other control lines, such as CRS (carrier sense) and COL (collision), are required for CSMA/CD specic functions.
idle sequences and data sequences exhibit identical power spectral densities; reception of an idle sequence can rapidly be distinguished from reception of data; scrambler state, pair A and pair B assignment, and temporal alignment and polarities of signals received on these pairs can easily be recovered from a received idle sequence. At the core of idle sequence generation and side-stream scrambling is a binary maximumlength shift-register (MLSR) sequence f pk g (see Appendix 3.A) of period 233 1. One new bit of this sequence is produced at every modulation interval. The transmitters in the master and slave transceivers generate the sequence f pk g using feedback polynomials g M .x/ D 1 C x 13 C x 33 and g S .x/ D 1 C x 20 C x 33 , respectively. The encoding operations
1267
are otherwise identical for the master and slave transceivers. From delayed elements f pk g four derived bits are obtained at each modulation interval as follows: x k D pk yk D pk a k D pk bk D p k
3 4 1 2
pk pk pk pk
8 6 5 12
(19.18)
where denotes modulo 2 addition. The sequences fx k g, fyk g, fak g, and fbk g represent shifted versions of f pk g, that differ from f pk g and from each other only by large delays. When observed in a constrained time window, the ve sequences appear as mutually uncorrelated sequences. Figures 19.13 and 19.14 illustrate the encoding process for the idle mode and data mode, respectively. Encoding is based in both cases on the generation of pairs of two-bit vectors .Xk ; Yk /, .Sa ; Sb /, and .Ta ; Tb /, and Gray-code mapping of .Ta ; Tb / into k k k k k k a b symbol pairs .Dk ; Dk /, where Dk 2 f 2; 1; 0; C1g, D a; b. The generation of these quantities is determined by the sequences fx k g, fyk g and f pk g, the even/odd state of the time index k (equal to 2n or 2n C 1), and the local receiver status. Finally, pairs of transmit a b A B symbols .ak ; ak / are obtained by scrambling the signs of .Dk ; Dk / with the sequences fak g and fbk g. In Figure 19.15, the symbol pairs transmitted in the idle and data modes are depicted as two-dimensional signal points.
Master/Slave bk MLSR sequence ak generator S S 1 1 S: 0 1 +1 1
n:odd
* x 2n+1
pk Xk Sk = Tka M
a
0 1
a Dk
{2,0,+2} {1,+1}
0 Y2n+1 Yk 1 Skb = Tkb M 0 [1,0] [1,1] 1 [0.1] [0,0] n:even even odd b Dk pk =0: a B k pk =1: akB {1,+1} {2,0,+2}
M( X k ) M( Yk )
{1,+1} {2,0}
1268
1 1 S: 0 1
2
1 +1
pk
xk
Tka
a Dk
yk
Skb
Tkb
a kA
2 1 0 1 2 1 2
0 1 2
2D symbol probabilities:
1/8
1/16
2D symbol probabilities:
Figure 19.15. Two-dimensional symbols sent during idle and data transmission.
A B We note that, in idle mode, if pk D 1 then symbols ak 2 Ax D f 1; C1g and ak 2 A 2 A and a B 2 A are transmitted. A y D f 2; 0; C2g are transmitted; if pk D 0 then ak y x k This property enables a receiver to recover a local replica of f pk g from the two received quinary symbol sequences. The associations of the two sequences with pair A and pair B can be checked, and a possible temporal shift between these sequences can be corrected. Idle sequences have the
b Dk
akB
akA
{2,1,0,+1,+2}
{2,1,0,+1,+2}
a kA
1/16
1/32
1/64
1269
further property that in every two-symbol interval with even and odd time indices k, two symbols 1, one symbol 0, and one symbol 2 occur. The signs depend on the receiver status of the transmitting transceiver and on the elements of the sequences fx k g, fyk g, fak g and fbk g. Once a receiver has recovered f pk g, these sequences are known, and correct signal polarities, the even/odd state of the time index k, and remote receiver status can be determined. In data mode, the two-bit vectors Sa and Sb are employed to side-stream scramble k k the data nibble bits. Compared to the idle mode, the signs of the transmitted symbols are scrambled with opposite polarity. In the event that detection of the delimiter marking transitions between idle mode and data mode fails due to noise, a receiver can nevertheless rapidly distinguish an idle sequence from a data sequence by inspecting the signs of the two received 2 symbols. As mentioned above, during the transmission of idle sequences one symbol 2, i.e. with absolute value equal to 2, occurs in every two-symbol interval. The previous description does not yet explain the generation of delimiters. A start-ofstream delimiter (SSD) indicates a transition from idle-sequence transmission to sending packet data. Similarly, an end-of-stream delimiter (ESD) marks a transition from sending packet data to idle sequence transmission. These delimiters consist of two consecutive A B A B symbol pairs .ak D 2; ak D 2/ and .akC1 D 2; akC1 D 0/. The signs of symbols 2 in a SSD and in an ESD are selected opposite to the signs normally used in the idle mode and data mode, respectively. The choice of these delimiters allows detection of mode transitions with increased robustness against noise.
19.2.1
The principal signal processing functions performed in a 100BASE-T2 transmitter and receiver are illustrated in Figure 19.16. The digital-to-analog and analog-to-digital converters operate synchronously, although possibly at different multiples of 25 MBaud symbol rate. Timing recovery from the received signals, as required in slave transceivers, is not shown. This function can be achieved, for example, by exploiting the strongly cyclostationary nature of the received signals (see Chapter 14).
1270
Figure 19.16. Principal signal processing functions performed in a 100BASE-T2 transceiver. [From [2] c 1997 IEEE.]
1271
0 -10 Power spectral density [dB] -20 -30 -40 -50 -60 -70
10
20 25 30
40
50 f [MHz]
60
70
80
90
100
Figure 19.17. Spectral template specied by the 100BASE-T2 standard for the power spectral density of transmit signals and achieved power spectral density for a particular transmitter implementation comprising a 5-tap digital transmit lter, 100 MHz D/A conversion, and a 3 order Butterworth analog transmit lter. [From [2], c 1997 IEEE.]
The use of forward equalizers with T =2-spaced coefcients serves two purposes. First, as illustrated in Section 8.4, equalization becomes essentially independent of the sampling phase. Second, when the received signals exhibit excess bandwidth, the superposition of spectral input-signal components at frequencies f and f 1=T , for 0 < f < 1=.2T /, in the T -sampled equalizer output signals, can mitigate the effects of synchronous interference and asynchronous disturbances, as shown in Appendix 19.A. Interference suppression achieved in this manner can be interpreted as a frequency diversity technique [5]. Inclusion of the optional cross-coupling feedforward and backward lters shown in Figure 19.16 signicantly enhances the capability of suppressing alien NEXT. This corresponds to adding space diversity at the expense of higher implementation complexity. Mathematical explanations for the ability to suppress synchronous and asynchronous interference with the cross-coupled forward equalizer structure are given in the Appendix 19.A. This structure permits the complete suppression of the alien NEXT interferences stemming from another 100BASE-T2 transceiver operating in the same multi-pair cable at identical clock rate. Alternatively, the interference from a single asynchronous source, e.g. alien NEXT from 10BASE-T2 transmission over an adjacent pair, can also be eliminated. The 100BASE-T2 standard does not provide the transmission of specic training sequences. Hence, for initial receiver-lter adjustments, blind adaptation algorithms must be employed. When the mean-square errors at the symbol-decision points reach sufciently low values, lter adaptation is continued in decision directed mode based on quinary symbol
1272
decisions. The lter coefcients can henceforth be continuously updated by the LMS algorithm to track slow variations of channel and interference characteristics. The 100BASE-T2 Task Force adopted a symbol-error probability target value of 10 10 that must not be exceeded under the worst-case channel attenuation and NEXT coupling conditions when two 100BASE-T2 links operate in a four-pair UTP-3 cable, that are illustrated in Figure 4.23. During the development of the standard, the performance of candidate 100BASE-T2 systems has been extensively investigated by computer simulation. For the scheme ultimately adopted, it was shown that by adopting time spans of 32T for the echo and self NEXT cancellers, 12T for the forward lters, and 10T for the feedback lters, the MSEs at the symbol-decision points remain consistently below a value corresponding to a symbol-error probability of 10 12 .
Note that the time span of an FIR lter is given in seconds by the product of the number of lter coefcients times the sampling period of the input signal. Transmission in a four-pair cable environment with suppression of alien NEXT from a similar transceiver is considered. Only the echo and self NEXT cancellers and forward equalizers will be compared. Updating of lter coefcients will be ignored. For dual duplex transmission, the modulation rate is 25 MBaud and signals are transmitted with about 100% excess bandwidth. Echo and self NEXT cancellation requires four FIR lters with time spans TC and input/output rates of 25 Msamples/s. For equalization and alien NEXT suppression, four forward FIR lters with time spans TE , an input rate of 50 Msamples/s and an output rate of 25 Msamples/s are needed. The modulation rate for mono duplex transmission is 1=T D 50 MBaud and signals are transmitted with no signicant excess bandwidth. Hence, both schemes transmit within a comparable bandwidth ( 25 MHz). For an obvious receiver structure that does not allow alien NEXT suppression, one self NEXT canceller with time span TC and input/output rates of 50 Msamples/s, and one equalizer with time span TE and input/output rates of 50 Msamples/s will be needed. However, for a fair comparison, a mono duplex receiver must have the capability to suppress alien NEXT from another mono duplex transmission. This can be achieved by receiving signals not only from the receive pair but also in the reverse direction of the transmit pair, and combining this signal via a second equalizer with the output of the rst equalizer. The additionally required equalizer exhibits the same complexity as the rst equalizer. The lter complexities for the two schemes are summarized in Table 19.1. As the required time spans of the echo and self NEXT cancellers and the forward equalizers are similar for
19. Bibliography
1273
dual duplex Echo and self NEXT cancellers 4 TC 25 1012 Forward equalizers 4 TE 50 25 1012
the two schemes, it can be concluded that the two schemes have the same implementation complexity. The arguments can be extended to the feedback lters. Finally, we note that with the lter time spans considered in the preceding section (TC D 32T , TE D 12T and TFb D 10T ), in a 100BASE-T2 receiver on the order of 1010 multiply-and-add operations/s need to be executed.
Bibliography
[1] G. Cherubini, S. Olcer, and G. Ungerboeck, A quaternary partial response class-IV transceiver for 125 Mbit/s data transmission over unshielded twisted-pair cables: principles of operation and VLSI realization, IEEE Journal on Selected Areas in Communications, vol. 13, pp. 16561669, Dec. 1995. [2] G. Cherubini, S. Olcer, G. Ungerboeck, J. Creigh, and S. K. Rao, 100BASE-T2: a new standard for 100 Mb/s ethernet transmission over voice-grade cables, IEEE Communications Magazine, vol. 35, pp. 115122, Nov. 1997. [3] G. Cherubini, S. Olcer, and G. Ungerboeck, Adaptive analog equalization and receiver front-end control for multilevel partial-response transmission over metallic cables, IEEE Trans. on Communications, vol. 44, pp. 675685, June 1996. [4] Supplement to carrier sense multiple access with collision detection (CSMA/CD) access method and physical layer specications: physical layer specication for 100 Mb/s operation on two pairs of Category 3 or better balanced twisted pair cable (100BASET2, Clause 32), Standard IEEE 802.3y, IEEE, Mar. 1997. [5] B. R. Petersen and D. D. Falconer, Minimum mean-square equalization in cyclostationary and stationary interferenceAnalysis and subscriber line calculations, IEEE Journal on Selected Areas in Communications, vol. 9, pp. 931940, Aug. 1991.
1274
Appendix 19.A
Interference suppression
Figure 19.18 illustrates the interference situations considered here. Equalization by linear forward lters only is assumed. Reception of 100BASE-T2 signals is disturbed either by alien NEXT from another synchronous 100BASE-T2 transmitter or by cross-talk from a single asynchronous source. Only one of these disturbances may be present. The symbol A B sequences fak R g and fak R g denote the sequences transmitted by the remote 100BASET2 transceiver, whereas fak T g and fak T g denote the sequences transmitted by an adjacent synchronous 100BASE-T2 transmitter. The spectrum S. f / of the asynchronous source may be aperiodic or exhibit a period different from 1=T . The functions H . f / represent the spectral responses of the signal or cross-talk paths from the respective sources to the inputs of the forward equalizer lters with transfer functions C A A . f / and C B A . f /. Because of 2=T sampling rate, these functions exhibit 2=T -periodicity. All signals and lter coefcients are real-valued. It is therefore sufcient to consider only frequencies f and f 1=T , for 0 < f < 1=.2T /. We will concentrate on the signals arriving at decision point DPA; the analysis for signals at DPB is similar. A Intersymbol-interference free reception of the symbol sequence fak R g and the suppresB sion of signal components stemming from fak R g at DPA require
c HA. f / C c A. f / C A c HA A0 B0
f f
1 T 1 T
CAA CB A
f f
1 T 1 T
D1 D0 (19.20)
c c HB . f / C B A . f / C
c HB
B (a) {a k T }
{a k T } HBA (f)
nxt nxt HAA (f) xt
(b) S(f)
{a k R }
CBA (f)
HBB (f)
nxt
HAB (f)
nxt
H xt (f)
SB
DPB
HB (f)
{a k R }
Figure 19.18. Cross-talk disturbance by: (a) alien NEXT from another synchronous 100BASET2 transmitter, (b) an asynchronous single source, for example, a 10BASE-T transmitter. [From [2], c 1997 IEEE.]
1275
To suppress alien NEXT from a 100BASE-T2 transmitter, two additional conditions must be met: X ` ` ` ` nxt nxt H A0 A f CAA f C H A0 B f CB A f D0 T T T T `D0;1 (19.21) X ` ` ` ` nxt nxt HB 0 A f CAA f C HB 0 B f CB A f D0 T T T T `D0;1 Alternatively, the additional conditions for the suppression of cross-talk caused by a single asynchronous source become
xt xt HS A . f / C A A . f / C HS B . f / C B A . f / D 0 1 1 1 xt C HS B f CB A f D0 T T T
xt HS A
1 T
CAA
(19.22)
Therefore in each case the interference is completely suppressed if for every frequency in the interval 0 < f < 1=.2T / the transfer function values C A A ( f ), C A A ( f .1=T /), C B A ( f ) and C B A . f .1=T // satisfy four linear equations. It will be highly unlikely that the cross-talk responses are such that the coefcient matrix of these equations becomes singular. Hence a solution will exist with high probability. In the absence of lter-length constraints, the T =2-spaced coefcients of these lters can be adjusted to achieve these transfer functions. For a practical implementation a trade-off between lter lengths and achieved interference suppression has to be made.
Index
Access methods, 523 Active device, 257 Adaptation gain, 168, 170, 177, 183185, 189, 212, 645, 703, 1086, 1100, 1104, 1120, 1124, 1250 Adaptive differential pulse code modulation (ADPCM), 341, 393, 398, 456, 1183 Advanced Mobile Phone Service (AMPS), 1180 Algorithm BahlCockeJelinekRaviv (BCJR), 668 BenvenisteGoursat, 1091 Contour (CA), 1093, 1095, 1098, 1102, 1113 DelsarteGenin, 110, 147 Fano, 917918 forward backward (FBA), 670, 915, 930, 939, 941, 948 Godard, 1101 Jacobi, 158 LindeBuzoGray (LBG), 424 least-mean-square (LMS), 173179, 186191, 205216, 628, 633, 638, 648649, 701, 722, 819, 1090, 1120, 1121, 1128, 1251, 1253, 1258, 1272 LempelZiv, 434 LevinsonDurbin, 145 Lloyd, 422 Max, 369 MuellerMuller, 1064 recursive least-squares (RLS), 197203, 217, 645, 722, 1180
Sato, 1090, 1096, 1100, 1106 stochastic gradient (SGA), 1086, 1113 stop-and-go, 1092 Viterbi (VA), 663, 677, 682, 686, 691, 823, 915, 921923, 968, 993, 1007, 1177, 1238, 1262 American National Standards Institute (ANSI), 1152 Analog-to-digital converter (ADC), 331, 338, 341, 385, 572573, 11891191, 1249 Antenna array, 226 directional, 296, 299, 811, 1165 gain, 296 isotropic, 296 Autocorrelation average, 1212 matrix, 6366, 132, 135, 149, 166, 420, 635, 719, 744 sequence, 50, 53, 86, 144, 401, 622, 655, 678 Automatic gain control (AGC), 1249 Automatic repeat query (ARQ), 807, 827 Autoregressive model (AR), 9194, 96101, 101, 115, 143, 197, 414 Autoregressive moving average model (ARMA), 90, 94 Bandwidth denition of, 29 excess, 764, 777, 1131, 1152, 1263, 1269
1278
Index
Bandwidth (continued) minimum, 458, 552, 559, 578, 590, 609, 771, 808, 1003 Bit error probability, 340, 456, 475, 485, 494, 499, 578, 613, 899, 919, 941, 954, 1212, 1224 Bit loading, 1002, 1152 Blind equalization, 1083 Block codes BoseChaudhuriHocquenghem (BCH), 878898 generator matrix, 836 generator polynomial, 864 Hamming, 872 low-density parity check (LDPC), 946955 non binary, 960965 parity-check matrix, 833 ReedSolomon, 885 simplex, 875877 syndrome decoding, 839 Calculus of variations, 731, 1000 Carrier sense multiple access (CSMA), 1164 Cancellation cross-talk, 1116, 1251 echo, 11161130, 1145, 1269 Cell radius, 1165, 1182 Cellular systems, 1171, 1182 Channel additive white Gaussian noise (AWGN), 326, 439, 458, 503506, 543, 713, 999, binary symmetric (BSC), 456, 571 Channel model, 251, 296, 309, 313, 316, 322, 658, 811, 1116 Channel capacity, 503, 999 Cholesky decomposition, 155 Circular convolution, 2123, 205, 242, 707 Code division multiple access (CDMA), 523, 795, 802, 810, 818, 1180 Code rate, 827, 828, 903, 1177, 1180, 1182, 1185
Codes block, 827, 899, 946, 990 BoseChaudhuriHocquenghem (BCH), 878898 channel, 827965 concatenated, 921, 924 convolutional, 900920 cyclic, 862898 forward error correction (FEC), 827 linear, 830, 968 low-density parity check (LDPC), 946955 ReedSolomon, 885 turbo, 924942 WalshHadamard, 536 Coding adaptive predictive (APC), 341, 401 adaptive transform (ATC), 341, 433 code excited linear predictive (CELP), 341, 416, 433, 578 linear predictive (LPC), 341, 414 residual excited linear predictive (RELP), 341, 415, 1226 subband (SBC), 341, 456 vector sum excited liner predictive (VSELP), 457, 1177 Coding by modelling, 413 Coherence time, 311, 314 Combining equal gain (EGC), 720 maximal ratio (MRC), 720, 815 optimum (OC), 721 selective, 719 switched, 720 Convolution, 13 Convolutional codes, 900920 catastrophic error propagation, 910 decoding algorithms, 912918 general description, 903 transfer function, 907 Correlation coefcient partial (PARCOR), 146
Index
1279
Correlogram, 86 Coset, 837841, 871, 964965, 977 Costas loop, 10401043 Criterion least squares (LS), 148, 197, 429 mean-square error (MSE), 166, 366 Cross-talk far-end (FEXT), 290, 1116 near-end (NEXT), 288289, 1116, 1251, 1269 Cut-off rate, 509 Cyclostationary process, 56, 1057 Decimation, 106, 236, 380, 766, 1126 Decoding iterative, 929, 939 sequential, 958 Delay spread rms, 307, 1177 Delta modulation (DM), 343, 404 adaptive, 343 continuously variable slope (CVSDM), 411 linear, 407 Demodulation coherent, 499 non coherent, 487, 806, 1213 Despreading, 795, 801 Detection decision feedback sequence (DFSE), 695 maximum likelihood (MLSD), 662, 969, 1262 multiuser, 820, 823, 1131 reduced state sequence (RSSE), 691696, 716 single-user, 818 threshold, 462, 474, 542, 555 Diagram state, 902, 909, 910 tree, 694, 901, 917, 974, 977 trellis, 664, 667, 677, 682, 903, 921, 968, 971, 981, 986, 990, 1014 Differential PCM (DPCM), 343, 385, 407
Digital European Cordless Telephone (DECT), 523, 1145, 1162, 11821185 Digital signal processor (DSP), 709 Digital subscriber line (DSL), 285, 11311142, 1147 asymmetric (ADSL), 11481150 high bit rate (HDSL), 523, 11471148 single line high speed (SHDSL), 919, 1148 very high speed (VDSL), 285, 523, 1115, 11311142 Digital-to-analog converter (DAC), 113, 331, 339, 408, 413 572, 1234, 1251 Distance Euclidean, 7, 420, 499, 682, 915, 967 free Euclidean, 919, 967, 986 free Hamming, 907 Hamming, 830, 875, 960 Distortion envelope delay, 319 Diversity frequency, 521, 1167, 1271 polarization, 522 space, 522, 1271 time, 522 Division multiplexing frequency (FDM), 753, 1175 time (TDM), 524, 1175 wavelength (WDM), 292 Doppler frequency, 318 shift, 303 spectrum, 311, 313, 315 spread, 304, 313, 1167 Downlink, 802, 819, 1156, 1175, 1183 Duplex transmission dual, 1154, 1263 full, 522, 1117, 11451148, 1251, 1264 half, 522, 1146, 1154 mono, 1154, 1264, 1272
1280
Index
Duplexing digital, 1149 frequency division (FDD), 523, 1146, 1175 time division (TDD), 522, 1184 Dynamic channel allocation (DCA), 1182 Dynamic channel selection (DCS), 1182 Echo cancellation, 225, 11161130 Envelope detector, 512 Equalization adaptive, 304, 645, 1018, 1083, 1252 fractionally spaced, 630, 642, 1269 self-training, 1083 Equalizer adaptive, 628, 1083, 1252 decision feedback (DFE), 595, 617, 635645, 649656, 695, 710, 741, 777, 823, 1010, 1131, 1270 decision feedback-zero forcing (DFE-ZF), 649, 657, 679, 695 fractionally spaced (FSE), 617, 630634, 642, 699, 703, 1107, 1270 linear (LE), 594, 619620, 627628 linear zero forcing (LE-ZF), 619 passband, 697 Estimation error, 129, 151, 159, 165, 174, 200, 213, 241, 622, 741, 1064, 1105 Estimate biased, 83 unbiased, 82 Estimator early-late, 1055 feedback, 1053 phasor, 1066 timing, 1055 Ethernet, 290, 11521155, 1263 European Telecommunications Standards Institute (ETSI), 415, 435, 1137, 1148, 1163, 1185 Excess MSE, 854
Extension eld, 890, Eye diagram, 562 Fading channel at, 305, 311, 718 frequency selective, 305, 311, 718 Rayleigh, 308, 518 Rice, 308 FCC, 1162, 1264 Fiber-to-the-curb architecture (FTTC), 1148 Fiber distributed data interface (FDDI), 1155 Filter allpass, 28 bank, 214, 433, 753755, 757773, 783, 818, 1152, 1190 Butterworth, 313 decimator, 110, 119126, 628, 754755 distributed arithmetic, 1121 nite impulse response (FIR), 25, 116, 129, 165, 399, 754, 1234, 1272 frequency response, 18, 543, 591, 760, 1028 highpass, 28 impulse response, 18, 542, 558, 754, 1005 innite impulse response (IIR), 25, 90, 136, 317, 622 integrate and dump, 1197 interpolator, 112126, 315, 754, 773, 1049, 1229 lattice, 146, 191, 204 loop, 1030, 1036, 1054, 1262 lowpass, 28 matched, 73, 462, 494, 567, 621, 698, 755, 799, 813, 1003, 1045, 1243 narrowband, 28 notch, 224 prediction error, 142, 386, 650, 679 transfer function, 16, 91 transversal, 165, 627, 11191120, 1255
Index
1281
whitened matched (WMF), 651, 1003, 1007 whitening (WF), 94, 651, 1005 Wiener, 129140, 165, 622 Filter bank critically sampled, 764767, 769 non critically sampled, 764769, 777 Finite eld, 844 Finite state machine (FSM), 663, 751, 970 Flexible precoding, 10181025, 1146 Free space path loss, 298 Frequency deviation, 44, 453, 1197, 1207, 1214 Frequency division duplexing (FDD), 523, 1146, 1175 Frequency reuse, 810, 1165, 1172 Front-end architectures, 1189 Function Bessel, 308, 511, 1056 Marcum, 529 saw-tooth, 1048 Galois eld, 844, 851 Gauss quadrature rule, 607 Gaussian random process, 68, 309, 503, 518 Gaussian random variable, 67, 314, 439 Generator matrix, 836, 865, 906 polynomial, 863, 869, 875, 905, 928 Geometric mean, 53, 1001 Global system for mobile communication (GSM), 523, 11721177 Gradient vector, 132, 150, 167, 1086 GramSchmidt orthonormalization procedure, 8 Granular error, 350, 354, 364 noise, 389, 408, 411 Graph, 907, 947, 952 Gray coding, 462, 531
Group, 830, 961, 977 Guard time, 1176, 1184 Hard input decoding, 913 Heaviside conditions, 32, 259 Householder transformation, 158 Hybrid ber/coax (HFC) networks, 11561160 IEEE 802 Working Group, 11521156, 11631164, 1266 Indoor environment, 1185 Inner product, 3 Integrated Services Digital Network (ISDN), 1119, 11471148, 1173 Intensity prole of multipath, 307 Interference cancellation, 1115 co-channel (CCI), 803 intersymbol (ISI), 456, 557, 604, 620, 655, 674, 749 multiuser (MUI), 803, 817 Interleaving, 913 International Telecommunications Union (ITU), 435, 1146, 1148 Interpolation, 31, 109, 116118, 339, 764, 1255, 1269 Jakes model, 313 Japanese Digital Cellular (JDC), 1171, 1180, 1235 Lagrange interpolation, 118 Lattice Gosset, 977, 991 Schlaei, 976, 991 Law A, 363 , 364 Leakage, 86, 1190 Likelihood function, 511, 674, 1051 Likelihood ratio, 442, 675 Limiter-discriminator, 1197 Line coding, 583601 Line codes biphase, 584
1282
Index
Line codes (continued) block, 585 dicode, 583, 591 duobinary, 591 modied duobinary, 592 NRZ, 583 RZ, 584 Line-of-sight (LOS), 295, 302 LMS algorithm for lattice lters, 191 in a transformed domain, 211 leaky, 187, 634 normalized, 189 sign, 187 Local area network (LAN), 290, 523, 1152, 1155 Local multipoint distribution service (LMDS), 1165 Log-likelihood function, 512, 677 Log-likelihood ratio, 676, 916, 934 Matrix circulant, 243, 776 diagonal, 20, 22, 155 generator, 836, 865, 906 Hadamard, 536 Hermitian, 65 inverse, 145 parity check, 833836, 840, 865, 870, 905, 906, 946950, 962965 Toeplitz, 63 triangular, 214 unitary, 66, 168, 182, 212 Vandermonde, 890 Maximum a posteriori (MAP) criterion, 160, 441, 661, 675, 677, 915, 921, 932, 948 Max-Log-MAP, 675, 677, 916, 921 Log-MAP, 677, 916, 952 Mean convergence of the, 178 Medium access control (MAC), 1156, 11631164 Message-passing decoding, 946, 953 Minimum function, 857861
Modulation amplitude (AM), 295, 461, 480, 539 amplitude and phase (AM-PM), 480 binary, 437, 487, 520 binary phase shift keying (BPSK), 450, 470, 477, 795, 919 biorthogonal, 493 carrierless AM/PM (CAP), 568570 continuous phase (CPM), 12461248 continuous phase FSK (CPFSK), 1217, 1219 differential PSK (DPSK), 474475 DMT, 770, 775, 781 double sideband (DSB), 41, 58, 608 DWMT, 782 FMT, 771, 777, 781 frequency (FM), 1197, 1213 frequency-shift keying (FSK), 452, 516, 519, 1207 Gaussian minimum-shift keying (GMSK), 12291243 index, 453, 1209 minimum-shift keying (MSK), 454, 12141228 multicarrier (MC), 753 offset QPSK (OQPSK), 1203, 1220 on-off keying (OOK), 510 orthogonal, 486 phase-shift keying (PSK), 465, 474, 995 pulse amplitude (PAM), 69, 72, 461464, 539543, 583, 1040 pulse duration (PDM), 464, 532, 534 pulse position (PPM), 464, 532 quadrature amplitude (QAM), 480485, 502, 544548, 611 quadrature PSK (QPSK), 472473 single sideband (SSB), 58, 499 trellis coded (TCM), 967998 vestigial sideband (VSB), 781 Moving average model (MA), 91, 94, 398
Index
1283
Multichannel multipoint distribution service (MMDS), 1165 Multipath, 299, 302, 307, 521, 717, 803, 811, 1167, 1243 Multiple access code division (CDMA), 524, 795, 802, 810, 818, 1180 frequency division (FDMA), 523, 1178 time division (TDMA), 524, 1158, 1175 Noble identities, 118, 759 Noise gure, 268, 270, 274 impulse, 782, 1158 shot, 265, 294, 326 temperature, 265, 273 thermal, 263264 Norm, 3 Numerically controlled oscillator (NCO), 1054 Nyquist criterion, 559, 562, 589 frequency, 462, 559, 616 pulse, 559, 801 OFDM, 753794 passband, 780 synchronization of, 779 systems, 769, 773, 780, 1002, 1128 Optical bers, 291294, 1148 Orthogonality principle, 134, 151 Oversampling, 404, 627, 630, 699, 1062 Parseval theorem, 15, 84 Partial response systems, 587, 1096, 1249 Per survivor processing (PSP), 695 Periodogram, 84 Welch, 85 Personal communication services (PCS), 810 Phase detector (PD), 1030 Phase deviation, 44, 322, 1193
Phase jitter, 321 Phase noise, 321, 697, 1189 Phase-locked loop (PLL), 326, 10291039, 1074, 1105, 1262 Pilot signal, 1027 Power amplier (HPA), 322, 1171 Power back-off, 11361141 Power delay prole, 310, 812, 1243 Power spectral density (PSD), 4662 Precoding, 596, 777, 599, 10081025 Prediction error, 141, 176, 192, 385, 399, 657 linear, 129, 140147, 414, 1180 optimum, 143, 392 Predictor backward, 140 forward, 140 linear, 140, 385, 391 optimum, 141, 399 Probability a posteriori, 160, 441, 661, 668, 931, 954 a priori, 440, 669 conditional, 159, 442, 510 transition, 932 Probability density function Gaussian, 567 Rayleigh, 308 Rice, 308, 518 Probability of error, 454, 483, 489, 494, 899 Processing gain, 809 Projection operator, 153 Pseudo-inverse matrix, 155157 Pseudo-noise (PN) sequences, 233238 CAZAC, 235 Gold, 236 maximal length, 233 Public switched telephone network (PSTN), 331, 1018, 1145 Pulse code modulation (PCM), 345, 357, 377, 385, 571 QPR-IV transceiver, 12491262 Quantization adaptive, 377
1284
Index
Quantization (continued) error, 345, 347, 350, 352, 387, 404, 575 scalar, 417, 420, 424 uniform, 355 vector, 417432 Quantizer adaptive, 377379, 381 non uniform, 358377 uniform, 346, 353 vector, 418, 421 Quotient group, 978 Radio link, 294318 Raised cosine pulse, 559561 Rake receiver, 811, 815, 1181 Rate bit, 331, 338, 340, 379, 406, 457, 504, 539, 578, 827830, 1002 sampling, 407, 555, 616, 626, 769 symbol, 458, 540, 754 Receiver direct conversion, 1190 superheterodyne, 1189 optimization, 731 Recovery carrier frequency, 1068 carrier phase, 1050, 1104 timing, 1046, 1053, 1060, 1074, 1252, 1269 Reection coefcient, 146, 278 Regenerative repeaters, 575581 Sampling theorem, 8, 30, 337, 627, 1046 Scattering, 295, 300, 310 Scrambling, 803, 1153, 1265 Self-training equalization 10831114 Set partitioning, 694, 973, 1014 Shadowing, 313 Shannon limit, 499, 504, 508, 924, 947 Signal analytic, 33, 38, 780, 804 bandwidth, 2829 baseband, 31, 36, 459 baseband equivalent, 3334, 549
complex envelope, 34 envelope, 43, 322 passband, 33, 697, 780 space, 14, 439, 990, 1008 Signal constellation, 12, 458 Signal-to-interference ratio (SIR), 798 Signal-to-noise ratio (SNR), 272, 449, 463, 469, 552, 626 Signal-to-quantization error ratio, 352, 354, 364, 387 Simplex transmission, 523 Simplex cyclic codes, 875878 Singular-value decomposition (SVD), 155 Slope overload distortion, 389 Soft output Viterbi algorithm (SOVA), 921924 Source coding, 433 Spectral efciency, 458, 460, 463, 468, 503, 782, 991 Spectral factorization, 53 Spread spectrum systems, 795826 applications of, 807 chip equalizer, 818 direct sequence, 795 frequency hopping, 804 symbol equalizer, 819 synchronization, 1074 Soft input decoding, 900, 914 Spreading techniques, 795 factor, 795, 1074 Standard ADSL, 1148 DECT, 1182 Ethernet, 1152, 1156, 12631273 FDDI, 1155 HDSL, 1147 HIPERLAN, 1185 IS-136, 1177 JDC, 1180 modem, 11461147 SHDSL, 1148 speech and audio, 434 Token Ring, 1155
Index
1285
VDSL, 1148 video, 435 Sufcient statistic, 440, 681, 816, 1007 Symbol error probability, 463, 469, 483, 508, 565, 604, 682 Synchronization, 10271082 Syndrome, 839841, 871, 889, 895, 965 System bandwidth, 265, 590, 1171 baseband equivalent, 549, 775 causal, 18 continuous time, 13 discrete time, 17, 556, 1003 identication, 239254 T1 carrier, 524 Tanner graph, 947 girth, 952 Telecommunications Industry Association (TIA), 286 Telephone channel, 318322, 697, 703 Token Ring, 1155 TomlinsonHarashima precoding, 1009 Transform D, 903 discrete cosine (DCT), 214215, 433, 790 discrete Fourier (DFT), 19, 206, 214, 433 fast Fourier (FFT), 20 Fourier, 1417
Hilbert, 3738 z, 18 Transmission lines, 274291 Trellis coded modulation (TCM), 967997 Union bound, 455 Universal mobile telecomunication service (UMTS), 1170, 1181 Uplink, 803, 1156, 1158, 1175 User code, 796, 803 Vector quantization (VQ), 417432 Vocoder, 341, 414 Voltage controlled oscillator (VCO), 1030, 1042, 1189, 1252 WienerKhintchine theorem, 46 Window Hann, 82 raised cosine or Hamming, 82 rectangular, 78, 82, 239 triangular or Bartlett, 82 Window closing, 86 Wireless local area networks (WLANs), 11621163, 1170, 1185 YuleWalker equations, 97 Zero-forcing equalizer LE-ZF, 619, 648649 DFE-ZF, 649, 651, 655