Beruflich Dokumente
Kultur Dokumente
lnvited Paper
0018-9219/85/o9oo-1349801.008 9 8 5 IEEE
than 2400 bits/s. Even 2400-bit/s modems nowoften not familiar with the field, is a simplified introduction to
incorporate this feature. intersymbol interference and transversal equalizers, and an
Voice-bandtelephonemodems maybe classified into overview of some practical adaptive equalizer structures. In
oneofthree categoriesbased on intendedapplication: the concluding sections of the first part, we briefly mention
namely,for two-wire public switched telephone network other related applications of adaptive filters (such as echo
(PSTN), four-wire point-to-point leased lines, and four-wire cancellation, noise cancellation, and prediction) and discuss
multipoint leased lines. PSTN modems can achieve 2400- past and present implementation approaches.
bit/s
two-wire
full-duplex transmission by sending 4 Before presenting the introductory material, however, it
bits/symbol and using frequency division to separate the seems appropriate to summarize the major areas of work in
signals in the two directions of transmission. Two-wire adaptive equalization, with reference to key papers and to
full-duplex modemsusingadaptiveechocancellation are sections of this articlewhere these topics are discussed.
n o w available for 2400- and W - b i t / s transmission. Adap- (The interested reader should refer to Lucky [59] and Price
tive echo cancellation in conjunction with coded modula- [83] for a comprehensive survey of the literature and exten-
tion will pave the way to %Oo-bit/s full-duplex operation sive bibliographies of work up to the early 1970s.) Unfor-
over two-wire PSTN circuits in the near future. At this time, tunately, use of some as yet undefined technical terms in
commercially available leased-line modems operateat rates the following paragraphs is unavoidable at this stage.
up to 16.8 kbits/s over conditioned point-to-point circuits, Nyquist’s telegraph transmission theory [I151 in 1928 laid
and up to 9.6 kbits/s over unconditioned multipoint cir- thefoundationfor pulsetransmission over band-limited
cuits. An adaptive equalizeris an essential component of all analog channels. In 1960, Widrow and Hoff [I091 presented
these modems. (See [27] for a historical note on voice-band a least mean-square (LMS) error adaptive filtering scheme
modem development.) which has beenthe workhorseadaptiveequalizational-
In radio and underseachannels, IS1 is due to multipath gorithm for the last decade and a half. However, research
propagation [87], [%I, which may be viewed as transmission on adaptive equalization of PAM systems in the early 1960s
through a group of channels with differing relative ampli- centered on the basic theory and structure of zero-forcing
tudes and delays. Adaptive equalizers are capable of cor- transversal or tapped-delay-line equalizers with symbol in-
recting for IS1 due to multipath in the same way as IS1 from tervaltapspacing [SS],
[56]. In parallel,thetheoryand
lineardistortionin telephone channels. In radio-channel structureof linear receiveand transmitfilters [116],[I171
applications, an array ofadaptive equalizerscanalso be weredevelopedwhichminimize mean-square errorfor
used to perform djversity combining andcancel inter- time-dispersiveadditive Gaussiannoise channels [31]. By
ference or jamming sources [6],[72]. One special require- the late I%Os, LMS adaptive equalizers had been described
ment of radio-channel equalizers is thattheybe able to and understood [33],[54],[&I]. It was recognized that over
track thetime-varyingfading characteristics typicallyen- highly dispersive channels even the best linear receiver falls
countered. The convergence rate oftheadaptation al- considerably short of the matched filter performance bound,
gorithm employed then becomes important during normal obtained by considering the reception of an isolated trans-
data transmission[87]. This is particularly true for3-kHz-wide mitted pulse [54]. Considerable research followed on the
ionospheric high-frequency (HF), 3- to 30-MHz, radio chan- theory of optimum nonlinear receiver structures under vari-
nels which suffer from severe time dispersion and relatively ous optimality criteria related to error probability [I], [59],
rapid time variation and fading. Adaptive equalization has [87]. This culminated in the development of the maximum-
also beenappliedtoslowlyfading tropospheric scatter likelihood sequenceestimator [24] usingtheViterbi al-
microwave digital radios, in the 4- to 11-GHz bands, at rates gorithm [25] and adaptive versions of such a receiver [17],
up to 200 Mbits/s [82]. [51],[61],[62], [MI, [89],[103]. Another branch of research
In the last decade there has been considerable interest in concentrated on a particularly simple suboptimum receiver
techniques for full-duplex data transmission atrates up to structure known as the decision-feedback equalizer [3], [4],
144 kbits/s over two-wire (nonloaded twisted-copper pair) [12], [32], [71], (831, [93]. Linear feedback, or infinite impulse
subscriber loops [2],[22],[69], [I&], [112]. Two competing response (IIR), adaptive filters [47] have not been applied as
schemes forachievingfull-duplextransmission are time- adaptive equalizers due to lack of guaranteed stability, lack
compressionmultiplex or burstmodeandadaptiveecho ofaquadratic performance surface, andaminorperfor-
cancellation. Some form of adaptive equalization is desir- mancegainover transversal equalizers [87]. As the ad-
able, if not indispensible, for these baseband modems due vantages of double-sideband suppressed-carrier quadrature
to a number of factors: high transmission rates specially for amplitudemodulation (QAM) over single-sideband (SSB)
theburst-mode scheme, attenuationdistortion based on and vestigial-sideband (VSB) modulation were recognized,
the desiredrange of subscriber loop lengthsand gauges, previously known PAM equalizers were extended to com-
and the presence of bridged taps, which cause additional plex-valued structures suitable for joint equalization of the
time dispersion. in-phase and quadraturesignals in a Q A M receiver [IS], [16],
The first part of this paper, intended primarily for those [49], [ a ] , [119]. Transversal and decision-feedback equalizers
QURESHI: A D A P T I V E E Q U A L I Z A T I O N 1351
A symbol x,,,, oneof L discreteamplitude levels, is synchronized t o the symbol rate. Fig. 5 shows the outline of
transmitted at instant mT throughthechannel,where T a trace (eye pattern) for a two-level or binary PAM system.
seconds is the signaling interval. The channel impulse re- If the channel satisfies the zero IS1 condition, there are only
sponse h( t ) i s shown in Fig. 3. The received signal r( t ) i s t w o distinct levels at the sampling time to. The eye is then
fully open and the peak distortion is zero. Peak distortion
(Fig. 5 ) is the IS1 that occurs when the data pattern is such
that all intersymbol interference terms add to produce the
maximum deviation from thedesired signal at the sampling
time.
The purpose of an equalizer, placed in the path of the
received signal, is to reduce the IS1 as much as possible to
maximize the probability of correct decisions.
Fig. 3. Channel impulse response.
the superposition of the impulse responses of the channel B. Linear Transversal Equalizers
to each transmittedsymbolandadditivewhite Gaussian
Amongthe many structures used forequalizationthe
noise n ( t )
simplest is the transversal (tapped-delay-lineornonrecur-
r( t) =Cxjh(t - /T) + n( t ) . sive) equalizershown in Fig. 6. In suchan equalizer, the
i current and past values r( t - n r ) of the received signal are
If we sample the received signal at instant k T + to, where linearlyweightedby equalizercoefficients(tap gains) c,
to accounts for the channel delay andsampler phase, we andsummedtoproducetheoutput.If the delays and
obtain tap-gain multipliers areanalog, the continuous output of
the equalizer z ( t ) is sampled at thesymbolrateand the
r( to +k ~ =) xkh( to) samples go to the decision device. In the commonly used
digitalimplementation, samples ofthereceived signal at
+ xjh( to + kT - j T ) + n( to + k T ) . thesymbol rate are stored in a digitalshift register (or
j# k
memory), and the equalizer output samples (sums of prod-
The first term on the right is the desired signal since it can +
ucts) z(to k r ) or zk are computeddigitally,onceper
be used to identifythe transmitted amplitude level. The last symbol according to
term is theadditive noise, whilethemiddle sum is the N-1
interferencefromneighboring symbols. Each interference zk = c,r( to + kT - nt)
term is proportional to a sample ofthechannelimpulse n-0
+
response h(fo ir) spaced a multiple iT of symbol inter-
where N is thenumberof equalizercoefficients,and to
vals Taway from to as shown in Fig. 3. The IS1 is zero if and
denotes sample timing.
+
only if h(fo ir) = 0, i # 0; that is, if the channel impulse
The equalizer coefficients c,, n = 0,l;. N - 1 may be
e ,
Fig. 4(a) and (d) shows the amplitude response of two C( f ) H ’ ( f ) = 1, If1 < 1/2T.
linear-phase low-pass filters: one an ideal filter with Nyquist Fromtheabove expression we see that an infinite-length
bandwidth and the other with odd (or vestigial) symmetry zero-IS1 equalizer is simply aninverse filter, which inverts
around 1/2T hertz. As illustrated in Fig. 4 (b) and (e), the thefolded-frequency response ofthechannel. A finite-
folded frequency response of ech filter satisfies Nyquist’s length ZF equalizer approximates this inverse.Suchan
firstcriterion. One class of linear-phase filters, which is inverse filter may excessively enhance noise at frequencies
commonly referred to in the literature [54], [87], [I131 and is where the folded channel spectrum has high attenuation.
widely used in practice [23], [118], is theraised-cosine This is undesirable,particularlyfor unconditionedtele-
family with cosine rolloff around 1/2T hertz. phone connections, which may have considerable attenua-
In practice, the effect of IS1 can be seen from a trace of tion distortion, and for radio channels, which may be sub-
the received signal on an oscilloscope with its time base ject to frequency-selective fades.
- M1 - 1t nn 1 M -1 M m
-1 M 0
I
P U I Q C B T V Z F ~
C. Automatic Synthesis the equalizer coefficients to reduce the sum of the squared
errors. The most popular equalizer adjustment method in-
Before
regular
data
transmission begins, automatic
volves updates to each tap gain during each symbol inter-
synthesis of the ZF or L M S equalizers for unknown chan-
val. Iterative solution of the coefficients of the equalizer is
nels, which involvestheiterativesolutionofoneof the
possible because the mean-square error (MSE) is a quadratic
above-mentioned sets of simultaneous equations, may be
function of the coefficients. The M S E may be envisioned as
carried out during a training period. (In certain applications,
an N-dimensional paraboloid (punch bowl) with a bottom
such as microwavedigitalradio systems, andremotesite
orminimum. The adjustment to each tapgain is in a
receivers in a multipoint telephone modem network [43],
direction opposite toan estimate of the gradient of theM S E
the adaptive equalizers are required to bootstrap in a deci-
with respect to that tap gain. The idea is to move the set of
sion-directed mode (see Section I-E) without the help of a
equalizercoefficients closer to theuniqueoptimum set
training sequence from the transmitter.) corresponding to the minimum M S E . This symbol-by-sym-
During the training period, a known signal is transmitted
bo1 proceduredevelopedby Widrow and Hoff [I091 is
and a synchronized version of thissignal is generated in the commonly referred to as the continual or stochastic update
receiver to acquireinformationaboutthechannelchar-
method because, instead of the true gradient of the mean-
acteristics. The training signal may consist of periodic iso-
square error,
lated pulses or a continuous sequence with a broad, uni-
form spectrumsuch as thewidely used maximum-length a€[ e: I / a c n ( k)
shift-register or pseudo-noise (PN) sequence [9],[54], [76],
a noisy but unbiased estimate
[118],[119]. Thelatter has the advantage of much greater
average power, and hence a larger received signal-to-noise ae:/ac,( k ) = 2ekr( to + kT - nT)
ratio (SNR) forthe same peak transmittedpower. The
is used. Thus the tap gains are updated according to
training sequence must be at least as long as the length of
the equalizer so thatthetransmitted signal spectrum is c,( k + 1) = c,( k ) - Ae,r( to + kT - n T ) ,
adequately densein the channelbandwidth to beequalized.
n = 0,1;.., N- 1
Given a synchronized version of the known training sig-
nal, a sequence of error signals ek = zk - x k can be com- where c,(k) is the nth tap gain at time & , e , is the error
puted at the equalizer output (Fig. 8), and used to adjust signal, and A is a positive adaptation constant or step size.
which is, on the average, 3 dB worse than the minimum tweenthetwo channels.The latter maybecaused by
asymmetry in the channel characteristics around the carrier demodulator andthe phase errorcomputationcircuit is
frequency. reduced to the delaythroughthedecisiondevice. Fast
The coefficients are adjusted to minimize the mean of phase jitter can betrackedmoreeffectively because the
the squared magnitude of the complex error signal, e ( & )= delay through the equalizer is eliminated from the phase
+
e,( k ) je,( k ) , where e, and e, are the differences between correction loop. The same advantage can be attained with a
z, and z,,and their desired values.The update method is basebandequalizer by putting a jitter-tracking loop after
similar to the one used for the PAM equalizer except that the equalizer.
all variables are complex-valued
G. Decision- Feedback Equalizers
c,( k + 1) = c , ( k ) - Ae,y*( f, + kT- nT),
We have discussed placements and adjustment methods
n = 0,1;.., N -1 forthe equalizer, butthe basic equalizerstructure has
where is the complexconjugateof y. Again, the use of remained a linear and nonrecursive filter. A simple nonlin-
complex notation allows the writing of this single concise ear equalizer [3], [4], [32], [71],which [93], is particularly
equation, rather than two separate equations involving four useful for channels with severe amplitude distortion, uses
real multiplications, which is what really has to be imple- decision feedback to cancel the interference from symbols
mented. which have alreadybeendetected. Fig. 12 showssuch a
The complex equalizer can also be used at passband [6], decision-feedback equalizer (DFE). The equalized signal is
[I61to equalize the receivedsignal before demodulation, as the sum of the outputs of the forward and feedback parts
shown in Fig. 11. Here the received signal is split into its of the equalizer. The forward part is like the linear transver-
in-phase and quadrature components by a pair of so-called sal equalizer discussed earlier.Decisionsmade onthe
phase-splitting filters, with identicalamplitude responses equalized signal are fed back via a second transversal filter.
and phase responses that differ by 90'. The complex pass- The basic idea is that if the value of the symbolsalready
band signal at the output of these filters is sampled at the detected are known (past decisions areassumed to be
symbol rate and applied to the equalizer delay line in the correct), then the IS1 contributed by these symbols can be
same way as atbaseband.The complex outputofthe canceledexactly, by subtracting past symbol values with
equalizer is demodulated, via multiplication by a complex appropriate weighting
from
theequalizer
output. The
exponential as shown in Fig. 11, before decisions are made weights are samples ofthetailofthe system impulse
and the complex error computed. Further, the error signal is response including the channel and the forward part of the
remodulated before it is used in the equalizer adjustment equalizer.
algorithm. The main advantage of implementingthe The forward and feedback coefficients may be adjusted
equalizer in the passband is that the delaybetweenthe simultaneously to minimize the MSE. The update equation
l"D
BIONK
-1
for the forward coefficients is the same as for the linear lows: The coefficients of a linear transversal equalizer are
equalizer. The feedback coefficients are adjusted according selected to force the combined channel and equalizer im-
to pulse response to approximate a unit pulse. In a DFE, the
b,(k+I)=b,(k)+be,k,_,, rn=l;.. ,M ability of the feedback section to cancel the ISI, because of
anumber of pastsymbols, allowsmorefreedom inthe
where R, is the kth symboldecision, b,(k) is the mth choice of the coefficients of the forward section. The com-
feedback coefficient at time k , and there are M feedback binedimpulse response ofthe channeland theforward
coefficients in all. The optimum L M S settings of b, rn = sectionmay have nonzero samples followingthemain
1.; e, M , are those that reduce the IS1 to zero, within the pulse. That is, theforwardsection of a DFE need not
span ofthe feedback part, in a mannersimilar to a ZF approximate the inverse of the channel characteristics, and
equalizer. Note that since the output of the feedback sec- so avoids excessive noiseenhancementandsensitivity to
tionofthe DFE is a weighted sum ofnoise-free past sampler phase.
decisions, the feedbackcoefficientsplaynopart in de- When a particular incorrect decision is fed back, the DFE
termining the noise power at the equalizer output. output reflects this errorduring the next few symbols as the
Given the same numberofoverallcoefficients, does a incorrect decision traverses the feedback delay line. Thus
D F E achieve less M S E than a linear equalizer? There is no there is agreater likelihoodof moreincorrectdecisions
definite answer to this question. The performance of each following the first one, i.e., error propagation. Fortunately,
typeof equalizer is influencedbytheparticularchannel the errorpropagation in a DFE is notcatastrophic. O n
characteristicsand sampler phase, as well as the actual typical channels, errors occur in short bursts that degrade
number of coefficients and the position of the reference or performance only slightly.
main tap ofthe equalizer.However,the DFE can com-
pensatefor amplitudedistortionwithout as muchnoise
H. Fractionally Spaced Equalizers
enhancement as a linear equalizer. The DFE performance is
also less sensitive to the sampler phase [94]. A fractionally spacedtransversal equalizer [6],[39], [45],
An intuitive explanationfor theseadvantages is as fol- [58], [71], [W],[I041 is shown in Fig. 13. The delay-line taps
%!LJ FILTER
11. RECEIVER STRUCTURES A N D THEIR STEADY-STATE PROPERTIES Thus hb( t ) is the impulse response of the positive frequency
partofthe channel spectralresponse translated to base-
A. Baseband Equivalent Model band.
The noise waveform n ( t ) in (1) is also complex-valued,
To set a common frameworkfordiscussing various re-
i.e.,
ceiverconfigurations,wedevelopa baseband equivalent
modelofa passbanddata transmission system. We start
from a generic QAM system (Fig. 9) since it can be used to
The receive low-pass filters (in Fig. 9) typically perform two
implement any linear modulation scheme.
functions: rejection of the "double-frequency" signal com-
The passband transmitted signalcan be compactly writ-
ponents,andnoise suppression.The latter function is
ten as
accomplished by further shaping the baseband signal spec-
trum. In the baseband equivalent model, only the first of
I
s( t ) = Re xx,,a( t - nT) exp( j2TIfct) I these functions of the receive filters is performed by g ( t )
L n 1 and absorbed in h ( t ) given in (2).For simplicity, the noise
where { x , , } is the complex sequence of data symbols with n( t ) in the baseband equivalent model is assumed to be
in-phase (real) and quadrature (imaginary) components, x , white with jointlyGaussian real and imaginary components.
and x ; , respectively, such that x , = x,, +
jx;,,, a ( t ) is the
transmitpulse shape, and fc is the carrier frequency.We 6. Optimum Receive Filter
shall assume that the baseband transmit spectrum A( f ) is Given the received signal y ( t ) , what is the bestreceive
+
band-limited to If1 < (1 u)/2Thertzwherethe rolloff filter? This question has been posed andanswered in differ-
factor u , between 0 and 1, determinesthe excess band- ent ways by numerousauthors. Here, we follow Forney's
width over theminimum If1 < 1/2T. (Notethat greater development [24]. He showed that thesequence of T-spaced
than100-percent excess bandwidth is sometimes used in samples, obtained at the correct timing phase, at the output
radio and baseband subscriber loop transmission systems.) of a matched filter is a set of sufficient statistics for estima-
The received signal is tion of the transmitted sequence { x , , } . Thus such a receive
r( t ) = S( r) hp( t ) + np( t ) filter is sufficient regardless ofthe(linear or nonlinear)
signal processing which follows the symbol-rate sampler.
where h,(t) is the passband channelimpulse response, For the baseband equivalent model derived in the previ-
n,(t) is "passband" Gaussiannoise, andtheoperator * oussection, the receivefiltermust have an impulsere-
represents convolution. The in-phase and quadrature out- sponse h*( - t ) , where the superscript * denotes complex
puts of the receive low-pass filters, ~ ( tand
) ~ ( t )may, be conjugate. The frequency response of this matched filter is
represented in complex notation as y ( t ) = K ( t ) & I t ) + H*(f), where H( f ) is the frequency response of the chan-
Y( t ) = g( t ) * 4 t ) exp (-j2nCt)l nel model h( t ) .
Ifthe data sequence { x , , } is uncorrelated withunit
where g ( t ) is the impulse response ofthereceive filter.
power, i.e.,
Assume that the receive filter completely rejects the dou-
ble-frequency signal componentsproducedbydemodula-
tion and centered around 2fc. Then the baseband received
thenthe signalspectrum at thematchedfilter output is
signal may be written as
IH( f)12, and the noisepowerspectrum is IH(f)12.After
y( t ) = x x , , h ( t - nT) + n( t ) (1) T-spaced sampling, the aliased or "folded" signal spectrum
n is
where shh( f) = x I H ( f - "/T)12, 0 < f < 1/T
n
is thecomplex-valuedimpulse response ofthe baseband and the noise power spectrum is N&&( f).
equivalent model (Fig. 17). The real-valued passband chan- If the transmissionmedium is ideal, then the baseband
nel impulse response h,(t) and the complex-valued base- signal spectrum H ( f ) at the matched filter input is deter-
band channel impulse response, h b ( t ) , are related accord- mined solely bythetransmitsignal-shaping filters. From
ing to the discussion in Section I-Ait is clear that if IS1 is to be avoid-
ed, the composite of the transmit filter and receive matched S h h ( f ) C ( f ) , O Q f Q 1/T. The optimum C ( f ) is one which
filter response must satisfy the Nyquist criterion, i.e., minimizes the MSE at its output. The MSE is given by
f ) = Ro, 0 Q f < 1/T
shh(
E = Tl'rD - Shh( f) c( f ) 1 2 + NOShh( f ) l c ( f)12 df (3)
where, in general,
wherethe first term is theresidual IS1 power,andthe
Ro = r/l/rshh( f ) df.
0 second term is the output-noise power. Differentiating the
integrand with respect to C ( f ) and equating the result to
Therefore, the overall Nyquist amplitude response IH( f)12 zero, oneobtainstheminimum MSE (MMSE) equalizer
must be equally divided between the transmit and receive frequency response
filters. For instance, each filter mayhavean amplitude
response which is a square root of a raised-cosine character- c( f , = I/[
NO + shh( f)]. (4)
istic [54], [113]. Forsuchan ideal additive white Gaussian Thus the best (MMSE) matched-filteredequalizedsignal
noise(AWGN)channel,thematchedfilter,symbol-rate spectrum is given by
sampler, and a memoryless detector comprise the optimum
receiver. shh( f ) / [ NO + Shh( f>l. (5)
Let us n o w considerthemoreinteresting case of an Substituting (4) in (3), one obtains the following expression
AWGN channel with linear distortion. Forsuch a channel, for the minimum MSE achievable by a linear receiver:
thesimple linear receiver describedabove is no longer
adequate. The symbol-rate sampled sequence, though still cmin (linear) = T/~/'N~/[N, + Shh( f ) ] d f . (6)
providing a set of sufficient statistics, now contains inter- 0
symbolinterference in addition to noise. The currentre- The receiver structure (Fig. 18) comprising a matched filter,
ceived symbol is distorted by a linear combination of past a symbol-rate sampler, an infinite-length T-spaced equalizer,
andfuturetransmitted symbols.Therefore, a memoryless andamemorylessdetector is referred to as theconven-
symbol-by-symbol detector is not optimum for estimating tional linear receiver.
thetransmitted sequence. Nonlinear receivers which at- if the equalizer response C ( f ) is designed to satisfy the
tempt to minimize somemeasure of error probability are zero-IS1 constraint then C( f ) = I/&( f).
the subject of Section 11-D, wherethe emphasis is on Theoverallfrequency response oftheoptimum zero-
techniques which combine nonlinear processingwith adap- forcing (ZF) linear receiver is given by H*( f ) / S h h ( f ) , which
tive filters. forces the IS1 at the receiver outputto zero.The MSE
It is instructive to first studylinear receivers, which at- achieved by such a receiver is
temDt to maximizesignal-to-noiseratio (SNR) (minimize
fless detection. €,,(linear) = T/l'T&,/shh( fd
) f. (7 )
0
cZF(Iinear) is alwaysgreater than or equal to cmin (linear)
C. LinearReceivers
because no consideration is given to the output-noise power
We begin by reviewing the conventional linear receiver in designing the ZF equalizer. At high signal-to-noise ratios
comprising a matched filter, a symbol-rate sampler, an in- the t w o equalizers are nearly equivalent.
finite-length symbol-spacedequalizer,and a memoryless 2) lnfinitefength Fractionally SpacedTransversal Filter:
detector. in thefollowing section
weshow
that the In this section, we derive an alternative form of an optimum
matched-filter, sampler, symbol-spaced equalizer combina- linear receiver.
tion is a special case of a more general infinite-length Let us start with the conventional receiver structure com-
fractionally spacedtransversal filter/equalizer. In fact, this prisingamatchedfilter, symbol-rate sampler, T-spaced
general filter maybeused as thereceivefilterfor any transversal equalizer, and a memoryless detector. As a first
receiver structure without loss of optimality. Sections 1 l - U step, linearitypermits us to interchangetheorderofthe
and ll-C4 present a contrast between practical forms of the T-spaced transversal equalizer and the symbol-rate sampler.
conventional and fractionally spaced receiver structures. Next, the composite response H*( f ) C ( f )of the matched
7) Matched Filterand lnfinitefength Symbol-Spaced filter in cascade with a T-spacedtransversal equalizer may
Equalizer: If further processing of the symbol-rate sampled be realized by a single continuous-time filter. Let us assume
sequence at the output of a matched receive filter is re- thatthereceived signal spectrum H( f ) = 0 outsidethe
stricted to be linear, this linear processor takes the general range If1 Q 1/27, 7 Q T, and the flat noise spectrum is also
form of a T-spaced infinite-length transversalor nonrecur- limitedtothe same bandby an idealantialiasing filter
sive equalizer followed by a memoryless detector [31], [54]. with cutoff frequency 1/27. The composite matched-filter
Let theperiodic frequency response of the transversal equalizercanthen be realized by an infinite-lengthcon-
equalizerbe C( f ) . The equalized signal spectrum is tinuous-time transversal filter with taps spaced at 7-second
H( f ) , If1 Q M/2KT The MSE achieved in this case is the same as €,,(linear)
0, M/ZKT < I f ( < M/2T derived earlier (7).
3) FixedFilter and Finite- Length Symbol-Spaced Equalizer:
The frequency response of the digital filter is H*( f)C( f ) , I f ( Theconventional MMSE linear receiver is impracticalfor
< M/ZKT. Thus if the output of this filter was produced at t w o reasons. First, constraints of finite length and computa-
the rate M/T, the output signal spectrum would be tional complexity must be imposed on the matched filter as
well as the T-spacedtransversal equalizer. Secondly, in
H( f ) H * ( f ) C ( f ) , ( f l < 1/2KT
most applications, it is impractical to design, beforehand, a
0, M/2KT< If1 Q M/2T. filter which is reasonably matched to the variety of received
When the filter output is produced at thesymbol rate, it signalspectraresultingfrom transmissionsover different
has the desired aliased signal spectrum channelsoratime-varyingchannel. Thus themostcom-
monly used receiver structure comprises a fixed filter, sym-
zIH(f - n/T)12C( f - n/T), 0 Q f Q 1/T. bo1 rate sampler,and finite-length T-spaced adaptive
n equalizer (Fig. 18). The fixed-filter response is either matched
Noting that C( f ) is periodic with period 1/T, this output- to thetransmitted signalshapeor is designed as a com-
signal spectrum is recognized as S,,(f)C(f), which is the promise equalizer which attempts to equalize the average
same as fortheconventional MMSE linearreceiver (5), of the class of line characteristics expected for the applica-
provided C( f) is selected according to (4). tion. For the present discussion, let us assume that the fixed
This simple development proves the important point that filter has animpulse response p ( t ) and a frequency re-
an infinite-length fractionally spaced digital transversal filter sponse P ( f ) . Then, the T-spaced sampled output ofthis
is at oncecapableofperformingthefunctionsofthe filter may be written as
matched filter and the T-spaced transversal equalizer of the
Uk =zXnq(kT- nT) -I- vk (10)
conventional linear receiver. n
Let us further show that the symbol-rate sampled outputs
of a fractionally spaced digital filter form a set of sufficient where
statistics for estimation of the transmitted sequence under 4 t ) = P(t)*h(t)
the following conditions. The digital filter with tap spacing
and
KT/M has the frequency response H*( f ) C ( f ) , If1 < M/2KT,
wherethe received signal spectrum H(f) is zerooutside vk = /n( t)p( kT- t) dt.
the band If1 < M/ZKT, C ( f ) is periodic with period 1/T,
and C ( f ) is information lossless. A sufficient condition for Denoting the N equalizercoefficients at time kT by the
C ( f ) to be information lossless is that C ( f ) is invertible, column vector c k , and the samples stored in the equalizer
i.e., C( f ) # 0, 0 Q f Q 1/T. However, it is only necessary delay line by the vector uk, the equalizer output is given by
that C( f)/S,,( f ) is invertible, i.e., C( f)/S,,( f ) # 0, 0 < f
Q 1/T. In words, this condition implies that C ( f ) may not zk = CLuk
Qeq( f ) C Q( f - n / T )
n
is the aliased power spectrumof p ( t ) . The minimum Fig. 19. Linear receiver based on a fractionally spaced trans-
achievable M S E is given by versal equalizer.
~=O,I,~,...,(NK/M)-I. (29)
and a memoryless threshold detector is sufficient to mini-
The frequency response of the optimum FSE approaches mize the probability of error. When amplitude distortion is
(8) as N + 00, and its MSE approaches cmin (linear) given present in the channel, a linear receive filter, e.g., FSE, can
in (6). reduce IS1 and output MSE by providing the output signal
As the noise becomes vanishinglysmall, an infinitely long spectrum
FSE has a set of zero eigenvalues. This implies that there are
an infinite numberofsolutionswhichproducethe same
minimum MSE. The nonunique nature of the infinite FSE is Thecorrespondingoutput error powerspectrum is No/
evidentfromthe fact thatwhenboth signal andnoise +
[ N o Shh(f)].Thus noise power is enhanced at those fre-
vanish in the frequency range 1/2T< If1 < M/2T, the in- quencies where Shh( f ) < R,. A memoryless detector oper-
WIYrra)
WmOL
aut%cE
It I I+
KyIw*R)
RQNE
*+$i
'I
'I
I' I,
I! I I I
----.
by (12), except that the summation over k now excludes the One important factor in the practical performance of D F E
set 1 Q k g N, (which is in the span of the feedback filter). structures of all types is the effect of error propagation on
Finally the final error probability. Assuming correct decisions, the
improvement in output signal-to-MSEratioprovidedby a
bo,, = QCopt.
D F E reduces the probability of occurrence of the first error.
Type 2: In this case, theforwardequalizercoefficients However, once an error occurs it tends to propagate due to
can be obtained from (11) independently of the predictor incorrect decision feedback. Bounds on the error multipli-
coefficients. Let Ck beanN,-element vector at time kT cation factor have been developed by Duttweiler et al. 1121
consisting of the N, most recent error signals before pre- and by Belfiore and Park [4].
diction at instants k , k - I;.., k - N, - 1, with the for- 2) Decision-Aided /SI Cancellation: The concept of de-
ward equalizercoefficients at theiroptimal values.Then, cision feedback of past data symbols to cancel intersymbol
the set of optimum predictor coefficients bk is the solution interference can theoretically be extended to include future
of the normal equations datasymbols. If all past andfuture datasymbols were
assumed to be known at the receiver, then given a perfect
E [ 6;-1&-7] bk = E [ 6;f?k]. (38) model of the IS1 process, all IS1 could be canceled exactly
without any noiseenhancement. Such a hypothetical re-
Type 3: This case is similar to Type 1, except thatthe
ceiver could, therefore, achieve the zero-IS1 matched-filter
forward filter structuredescribed in Section ll-C4 is used.
bound on performance. In practice,theconceptofdeci-
The set of optimum forward coefficients can be obtained
sion-aided IS1 cancellation can beimplementedbyusing
using
tentative decisions and some finite delay at the receiver. In
cop, = k ' a the absence of tentative decision errors, the IS1 due to these
finite number of future data symbols (as well as past data
wherethematrix A has elementssimilar to A,,, givenby
symbols) can be canceled exactly before final receiver deci-
(26), except that the summation over k now excludes the
sions are made. Such a receiver structure with a two-step
set 1 Q K < N,. The optimum feedbackcoefficients are
decision process was proposedby Proakis[85].Recently,
given by
Gersho and Lim [34] observed that the MSE performance of
bop, = HCopt this receiver structure could be improved considerably by
an adaptive matched filter in the path of the receivedsignal
where H is an N, X N matrix with elements h,,j = h( iT -
prior to IS1 cancellation. In fact, the zero-IS1 matched-filter
jKT/M).
bound on M S E could be achieved in thelimit assuming
Type 4: Thistypeof D F E is similar to Type 2. The
correct tentative decisions. A general theoretical treatment
optimum forward coefficients given by (25) still apply and
is available i n [ a ] .
the predictor coefficients can be obtained by solving a set
Consider the block diagram of Fig. 23 where correct data
of normal equations similar to that given in (38).
symbols are assumed to be known to the canceler. Let the
forwardfilter have N coefficientsfractionally spacedat
For a direct comparison of the performance of the four
KT/M-second intervals. Then the final output of the struc-
types of DFE structures, resort must be made to numerical
ture is given by
solutionfor particularchannel characteristics, numberof
equalizer coefficients, etc. One general comment that can zk = c:yk - q x k
bemade is thatfor an equalnumberofforwardand
feedback coefficients, the D F E structures of Type 1 and 3 where ck and yk are the forward filter coefficient and input
will always achieve an output M S E at least as low or lower vectors, respectively, with N elements as defined earlier in
thanthe M S E achieved by structuresof Type 2 and 4, Section ll-C2, bk is the vector of the canceler coefficients
respectively.This is true because unlike Type 1 and 3 spaced T seconds apart, and xk is the vector of data sym-
structures,independentsolutionoftheforwardequalizer bols. Each of these vectors is of length N, +
N,. The data
and the feedback predictor coefficientsin Type 2 and 4 D F E vector xk has elements xk+ N,, . . . x k + ' , x ~ - ~ .,.,. x ~ - ~ , .
structures does not in general guarantee joint minimization Notethatthe current data symbol xk is omitted. The
of the final MSE. canceler has N1 noncausal and N, causal coefficients
RECEIVE FILTER
DELAYED
FINAL
MClSlONS
ERROR
SMUEMCE
obtain the powerspectrum of the combined errorsequence to two or three terms [17], [30], it invariably takes the form
at the input to the Viterbi algorithm. The error sequence is of one or the other familiar class of partial response systems
found to be white with mean-squared value No/( R, +
No), [48] with a null at one or both band edges. As mentioned in
which is equal to the matched-filter bound. Note that the Section 11-D, the Viterbi algorithm is able to recover most of
Gaussiannoisecomponentofthe errorsequence is ap- the nearly3-dB loss which otherwise results from use of
proximately white at a moderately high SNR. However, the bandwidth-efficient partial-response systems.
total errorsequence whilewhite is not Gaussian due to Several methodsofjointlyoptimizing thefractionally
residual ISI. spaced receive filter and thetruncated DIRare available
Let us select the DIR such that G ( f ) is the causal factor whichminimizethe M S E at theinputtotheViterbi al-
of the power spectrum given in (40). Then the resulting FSE gorithm (VA). These methods differ in the formofcon-
(39) may be recognized as the optimum forward filter of an straint [17], [ a ] on the DIR which is necessary in this
infinite-length conventional LMS DFE. At moderately high optimization process to excludetheselectionofthe null
SNR this FSE approaches the WMF. An MLSE receiver using DIR corresponding to no transmission through the channel.
such a receive filter is shown in Fig. 24. This receiver can The general form of such a receiver is shown in Fig. 25.
easily be made adaptive by updating both the FSE and the One such constraint is to restrict the DIR to be causal and
feedback filter coefficients to jointly minimize the mean- to restrict the first coefficient of the DIR to be unity. In this
squaredvalue of the error sequence.The feedback filter case, the delay LT in Fig. 25 is equal to the delay through
coefficients also provide all but the first DIR coefficient for the VA and the first coefficient of { 4 ) is constrained to be
use bytheViterbialgorithm. The first DIR coefficient is unity.
assumed to be unity. If the causality constraint is removed (as in [34] for deci-
Fredericsson [30] points out the difficulty of obtaining a sion-aided IS1 cancellation)butthereference(orcenter)
general explicit solution for the optimum truncated DIR of coefficient of the DIR is constrained to unity, another form
a specified finite length. However, when the DIR is limited of the receivershown in Fig. 25 is obtained [103]. In this
d j k= d, n
k-1
n-0
(1 - A n x i ) , i = 0,1;.., N -1
Three important results can be derived [38] using this line
of analysis [I021 and eigenvalue bounds [38]:
1 ) The L M S algorithm is stable if the step size A is in the
corresponding to (47). Observe that, i f the eigenvalues are range
known beforehand, complete convergence can be obtained
in N steps by using a variablestep size providedthe N 0 < A < 2 / ( NpX) (53)
values of the step size are selected such that A n = l / X n ,
n = 0 , l ; . ., N - 1 . Thuseach step reduces oneofthe
x
where is the average eigenvalue of A and is equal to the
average signal power at the equalizer input, defined accord-
tap-gain deviation components to zero. ing to E [ y : g ] / N . When A satisfies (53), all eigenvalues of
The second method of obtaining fast convergence is to the matrix M in (52) are less than one in magnitude permit-
replacethe scalar A in (44) bytheprecomputed inverse ting mean-square coefficient deviations in ( 5 2 ) to converge.
matrix A - l . This is tantamount to direct noniterative solu- 2 ) The excess M S E follows the recursive relationship
tion, since regardless of initial conditions or the eigenvalue
spread of A , or the equalizer size N , optimum solution is cAk+l = [ I - 2Ax + A2Np(x)'] C A k + A2cmlnNp(X)*.
obtained in one step.
(54)
If A is selected to minimize the excess MSE ateach itera-
B. LMS Gradient Algorithm tion, we would obtain
In practice, the channel characteristics are notknown ' k I 'Ak/('rnin + '&k)( Npx).
beforehand. Therefore, the gradient of the M S E cannot be Initially cAk x=-cmln, so that fastest convergence is obtained
determined exactly and must be estimated from the noisy with an initial step size
receivedsignal. The LMS gradientalgorithm [I091 is ob-
tainedfromthe deterministicgradientalgorithm (44) by A,N =
p XI /)(. (55)
replacing the gradient Note that A. is half as large as the maximum permissible
step size for stable operation.
3 ) The steady-state excess M S E can be determined from
by i t s unbiased but noisyestimate yf (y,'ck - x k ) . The (54) to be
equalizercoefficients are adjustedonce in every symbol
interval according to c,, = AemlnNpx/(2 - A NpX).
(56)
As mentioned earlier, only about K/M of the eigenvalues where (X/€) can be recognized as thedesiredequalizer
of the correlation matrix A are significant for a KT/M FSE. input-power-to-output-MSE ratio.
For the sameaveragesignal power at the equalizer input, Asan example,consider a 32-tap T-spaced equalizer
equal to the average eigenvalue x,thesignificanteigen- (N = 32) for a well-behaved channel ( p = 1). Select y =
0.25 (corresponding to a I-dB increase in output M S E over
values are generally M/K times larger for an N-coefficient
K T / M FSE compared with the eigenvalues for an N-coeffi- ),E
,, andadesiredequalizer output-signal-power-to-MSE
cient T-spaced equalizer. Thus the eigenvalue ratio p for an ratio of 24 dB (adequate for 9.6-kbit/s transmission). Let the
FSE should be computed only overthesignificanteigen- input-to-output power scale factor for this equalizer be 2,
values of A . Note that A and p for an FSE are independent so that = 1027/20. Solving (59) for B, we find that
of timing phase and channel envelope delay characteristics, 12-bit precision is required.
and p = 1 whenthe unequalizedamplitude shape is The required coefficient precision increases by 1 bit for
square-root of Nyquist. each doubling of the number of coefficients, and for each
For well-behaved channels ( p approximately I), cAk for
6-dB reduction in desired output MSE. However, for a given
an N-coefficient KT/M FSE, using the best initial step size c m i n ,each 6-dB reduction in excess MSE E,, requires a 2-bit
given by (55), initially convergesfaster as (1 - M / K N p increase in the required coefficient precision. This becomes
compared with (1 - l/N)k for an N-coefficient Tequalizer the limiting factor in some adaptive filters, e.g.,an echo
usingthe samebest initial step size.Conversely,
an canceler, which must track slow variations in system param-
MN/K-coefficient FSE, using a step size K / M times as eters inthe presence ofa large uncorrelatedinterfering
large, generally exhibits the same behavior with respect to signal [108].
the convergence and steady-state value of excess M S E as an Note that the precision requirement imposed by the L M S
N-coefficient T-spaced equalizer. gradient adaptation algorithm is significantly more stringent
As a rule of thumb, the symbol-by-symbol L M S gradient than a precision estimate based on quantization noise due
algorithmwiththe best initial A leads to a reductionof toroundoffincomputing the sum ofproductsforthe
about 20 dB in M S E forwell-behavedchannels ( p ap- equalizer output. If each product is rounded individually to
proximately 1) in about five times the time span cq of the B bits and then summed, the variance ofthis roundoff
noise is N2-”/3. Assumingan equalizer output signal
equalizer. At this time it is desirable to reduce A by a factor
of 2 for the next 5 cq seconds to permit finer tuning of the power of 1/6, the signal-to-output roundoff noise ratio is
2”/2N, which is 54 dB for B = 1 2 and N = 32:30dB
equalizer coefficient. Further reduction in excess MSE can
be obtained by reducing A to its steady-state value accord- greaterthanthedesired 24-dB signal-to-MSEratio in our
ing to (57). (For distorted channels, an effective time span example.
of p T , , should be substituted in the above discussion.) Roundoff in thecoefficientupdate process, which has
been analyzed in [7], is another source ofquantization
noise in adaptive filters. This roundoff causes deviation of
thecoefficientsfromthe values they take wheninfinite
C. Digital Precision Considerations
precisionarithmetic is used.The M S E E, contributedby
The above discussion maysuggest that it is desirable to coefficient roundoff [7] is approximated by E, = N2T2’/6A.
continue to reduce A in order to reduce the excess M S E to Againusing an output signal power value of1/6,the
zero in the steady state. However, this is not advisable in a signal-to-coefficient roundoff noise ratio is A2”/N. For our
practical limited precisiondigitalimplementationof the example, we obtain A = 0.0375 from (59) using y = 0.25,
adaptive equalizer. Observe [36] that as A is reduced, the N = 32, p = 1, and X = 1/3. The signal-to-coefficient
coefficient correction terms in (50), on the average, become roundoff noiseratio is 43dB for B = 12, which is 19 dB
smaller than half the least significant bit of the coefficient; greater than the desired 24-dB signal-to-MSE ratio, suggest-
adaptationstallsandthe M S E levels off. If A is reduced ing that the effect of coefficient roundoff is insignificant.
the scaler A in (SI), such that each element Ai of this matrix - x,$ = 8,k
is the inverse of the corresponding element X i of A . Trans- N n-0
forming diag(Ai) back totheoriginalcoordinate system, we have
we obtain the orthogonalized L M S update algorithm
V,* Vk = NAP and V,*xk = Nap.
ck+l = =k - W e k . (63) The elements of the matrix ,A
, and vector a, are given by
As in Chang’s scheme [8], the best value of the weighting N-1
matrix P is given by a,,j= h*( n T - iT) h( n T - jT)
n-0
P = V*’diag( Ai)V = V*’A-’V = A-’
and
where the columns of the diagonalizing matrix Vare eigen-
vectors of A (see (30)). In practice, A is not known before- a;= h*( - iT)
hand, therefore, P can only approximate A-’. For instance, where h( nT), n = 0,l; .
., N - 1, are the T-spaced samples
in partial-response systems wherethedominant spectral of the periodic channel response to a sequence of periodic
shape is known beforehand, we can use P = S1, where S impulses spaced N T secondsapart.Thus the periodic up-
is the covariancematrixofthepartial-responseshaping date algorithm may be written as
filter.
A practical advantage of this algorithm (63) over Chang’s
ck+1 = ( 1 - ANA,)ck + Ahfa,. (66)
structure is that the weighting matrix is in the path of the Notethesimilarityof (66) to thedeterministicgradient
tap-gaincorrectionsratherthanthereceivedsignal. The algorithm (44).However, in this case the matrix A, is not
computation required, therefore, need not be carried out to only Toeplitz but also circulant, i.e.,each row of A ,, is a
as much accuracy. circular shift of another row. The elements a c j of the i t h
As weshall see in Section IV-C,the fastest converging row are the coefficients R i - , of the periodjc autocorrelation
algorithms are obtained when P is continually adjusted to function of the channel. The final solution to thedifference
do the best job of orthogonalizing the tap-gain corrections. equation (66) is the “optimum” set ofperiodicequalizer
coefficients: cpopt = Ai's,. Substituting in (a),we obtain
12
N-1
c,( k + 1) = c,( k) -A ek,+jy*( kNT + j T - nT), =IFH(i/NT- k/T) , i=O,I;..,N-l.
j-0
n = 0,1;.., N - 1. (68)
Let us denote the error sequence during the kth period by Using these results, it can be shown that in the absence of
QURESHI: 1377
noisetheperfectperiodicequalizer has a frequencyre- Here
sponse equal to the inverse of the folded channel spectrum
R , = ( y l Y , * , ,n
, )=/ N
O, I ; * * , N - l
at N uniformly spaced discrete frequencies
C(n/NT) = 1 H(n/NT- k/T), n = 0,1;.., N - 1. are the coefficients of the periodic autocorrelation function
/ k ofthe equalizer input(definedearlier in terms ofthe
The excess M S E after the kth update is given by periodic impulse response of the channel). When R, = 0
for n > 1 , corresponding to an idealNyquistchannel, all
C d k = eThdk roots of (71) are equal to the Nth rootsof (1 - ANR,).
N-1 Thus perfect equalization is obtained after N updates with
= XPi(l - A N h , j ) z k ~ d , ~ 2 . A = I / ( NR,).
i-0 Let X,, be the maximum magnitude root of (71). Then
If the initial coefficients are zero, then Idjolz= I&, and the coefficient deviations after every N stochastic updates
are reduced in magnitude according to (70) provided IX%,I
< 1 . Moreover,fromthemaximummodulusprincipleof
holomorphic functions [92] we have the condition that
I N-1 .
.. I
Each component of this sum converges provided 0 < A <
2 / ( NXpmax).Fastest convergence is obtainedwhen A = IAN,,,( d max R,exp( -j27rni/N)
0s i < N-1
2/[N(Xpmax + A p m i n I l .
2) Stochastic Update: So far we have examinedthe or
convergenceproperties of theperiodicupdate or LMS
steepest descent algorithm with averaging. It is more com-
mon, and as we shall see, morebeneficial to use the where X,, are the eigenvalues of A, in the periodic update
continualor stochasticupdatemethod,where all coeffi- method. Since (72) holds with equality only when R , = 0
cients are adjusted in each symbolintervalaccording to for n 2 1, we reach the important conclusion for periodic
(50). Proceeding as in Section 111-6 and noting that in the trainingthatforthe idealNyquistchannelthestochastic
absence of noise ekopt= 0 for a periodic input, we obtain and averaged update algorithms converge equally fast, but
for all other channels the stochastic update algorithm re-
(ck+l - cpopt) = (I- Afiyl)(ck - Cpopt)
sults in faster convergence. This behavior has also been
Let us define an N X N matrix observed for equalizerconvergence in the presence of
random data [71].
Bk (1- Afiyl).
In the presence of noise,anexactexpression forthe
Note thatsince yk is periodic with period N , B, is a excess MSE for the stochastic update algorithm is difficult
circulant matrix and to derive for periodic training sequences. However, assum-
ing zero initialcoefficients the following expression isa
Bk+1 u-’Bku good approximation toresults obtained in practice for mod-
where the N X N cyclic shift matrix U is of the form erately high SNR:
ii
N-1
0 1 0 . . 0 chkN = [(I - + A 2 N h : ~ p m i n ] (73)
U= [0 1
0
0
0 0 . .
Ij.
i-0
adaptationalgorithm canbe derived as a solution to the The parameter 6 is selected as a small positive number to
exact least squares problem without any statistical assump- ensure that A, is nonsingular. The matrix A,, and vector a,
tions. An advantage of this approach is that the “shifting are akin to the statistical autocorrelation matrix and cross-
property”previously used for fastrecursiveleastsquares correlation vector encountered in the L M S analysis. How-
identification algorithms [73]can be applied to the equalizer ever, in this case A, is not a Toeplitzmatrix even for
adaptationalgorithm. This resulted in theso-called fast T-spaced equalizers.
Kalman algorithm [20] which requires on theorderof N It can beshown [87] thatgiven c,,-~, thecoefficient
operations per iteration. vector for time n can be generated recursively accordingto
A third class of recursive least squares algorithms known
as adaptive lattice algorithms [I271 were first described for
adaptive identification in [74] and for adaptive equalization where e, = x , - y;c,,-, is the equalizer output error and
in [63], [%I, and [97]. Like the fast Kalman algorithm, adap-
k, = A i l g
tive lattice algorithms are recursive in time, requiring of the
orderof N operations per iteration.However,unlikethe is the Kalmangainvector. Thepresence ofthe inverse
Kalmanalgorithms,adaptivelatticealgorithms are order- estimated covariance matrix in (81) explains the insensitivity
recursive. That is, the number of equalizer coefficients (and of the rate of convergenceofthe RLS algorithms tothe
the corresponding lattice filtersections) can be increased to channel characteristics.
+
N 1 without affecting the already computed parameters In the Kalman algorithm [41], the inverse matrix P, = A i 1
oftheNth-order equalizer.Lowsensitivityofthelattice and the Kalmangainvector are computedrecursively
coefficients to numericalperturbations is
a further ad- according t o
vantage.
In the remainder of this section, we shall briefly review
the leastsquare criterionand its variants, introducethe and
“shifting property” and the structure of the fast Kalman and
adaptive lattice algorithms, and summarize some important Pn = [ E-1 - kn~,T~n-l]/w. (82)
results, complexity estimates, and stability considerations. The order of N2complexity of this algorithm is due to the
1) TheLeastSquares Criterion: The performanceindex
explicit recursivecomputationof P,. This computation is
for recursive least squares (RLS) algorithms is expressed in
also susceptible to roundoff noise.
terms of a time average instead of a statistical or ensemble
Two other variations of the prewindowed and exponen-
average as in L M S algorithms. The RLS equalizer adaptation
tiallywindowed leastsquares criteria are the“growing
algorithm is required to generate the N-coefficient vector
memory” covariance and “sliding window” covariance per-
c, at time n which minimizes the sum of all squared errors
formance indices defined as
as if c, were used over all the past received signals, i.e., c,
minimizes
n Wk-,lXk - y:c,12
k-N-I
(77)
k-0 and
n
This leads to theso-called prewindowed RLS algorithm,
wheretheinput samples yk areassumed to be zerofor Ixk - YkTcn12
k=n-L+1
k < 0.
In order topermit trackingofslow time variations,a respectively,where N is thenumberof equalizer coeffi-
decay factor w with a value slightly less than unity may be cients and L is the fixed length of the sliding window. The
introduced. The resulting exponentially windowed RLS al- sliding window methodis equivalent to a block least squares
gorithm minimizes approach, where the block is shifted by one sample and a
n new optimum coefficient vector is determined each sample
time.
k-0 2) TheFast Kalman Algorithm: Consider the input vec-
tor Y,,-~ at time n - 1 for a T-spaced equalizer of length N.
The minimizing vector c, is the solution of the discrete-
The vector y, at time n is obtained by shifting the elements
time Wiener-Hopf equation obtainedby setting the deriva-
of Y,,-~ by one,discardingtheoldest sample y,-,,,, and
tive of (78) with respect to c, to zero.Thus we have the
adding a new sample y,. This shifting property is exploited
solution
byusing leastsquare linearprediction. Thusan efficient
c, = A i 1 a, recursive algorithm can bederived [20] forupdatingthe
Kalman gain vector k , without explicit computation of the
where the N X N estimated covariance matrix is given by
inverse matrix P,,. Here we shall briefly examine the role of
n
forwardandbackwardpredictioninthe fast Kalmanal-
k-0
gorithm. See [20] for a more detailed derivation generalized
to fractionally spaced and decision-feedback equalizers.
andtheN-element estimatedcross-correlationvector is Let F,,-l be a vector of N forward predictor coefficients
QURESHI: 1381
which minimizes the weighted sum of squares of the for- The transformed received signal components are then lin-
ward prediction error f, between the new input sample y, earlyweightedby a set of equalizer coefficientsand
and a prediction based on the vector Y , - ~ . That is, Fn-l summed to produce the equalizer output. We shall briefly
minimizes review the gradient [%] and least squares [87], [97] forms of
n adaptivelatticealgorithmsfor linear T-spaced complex
W"-klfk12 equalizers. See [78] and [52] for generalization of the least
k-0 squares AL algorithm to fractionally spaced and decision-
where feedback equalizers.
The structure of an adaptive lattice gradient equalizer is
fk Yk - 6 -TI Y k - 1 . (83)
shown in Fig. 26. An N-coefficient equalizer uses N - 1
The leastsquares forwardpredictorcoefficients can be
updated recursively according to ae*OvLass
F, = Fn-l + k,-,f,. (W
Similarly, the vector B,,-l of N backward predictor coeffi-
cients permits prediction of the old discarded sample yndN
given the vector y,. Thus we have the backward error
..+
bn = Yn-N - B:-lYn (85)
and the update equation
B, = B,-, + k,b,
The updated Kalman gain vector k,, which is notyet
available, can be obtained as follows. Define
C Yn - CYn-1 (86)
as the error between y, andits prediction based on the
updated forward predictor. Let Fig. 26. Gradient adaptive lattice equalizer.
E, = WE,-, + C*f, (87)
latticefilter stages.Each symbol interval a new received
betheestimatedexponentiallyweighted squared predic- sample y ( n ) enters stage 1. The mth stage produces t w o
tion error. Then the augmented or extended Kalman gain signals f,(n) and b,(n) which are used as inputs by stage
+
vector with N 1 elements is given by +
m 1. These signals correspond to the forward and back-
ward prediction errors,respectively, of mth-order forward
andbackwardlinear LMS predictors. The two predictors
have identical so-called reflectioncoefficients k , forthe
mth stage. Attime n, theprediction errorsare updated
where the dashed lines indicate partitions of the vector k,. according t o
Finally, theupdated backwardpredictor vector and the
Kalman gain vector are given by the recursive relationships 6(.) = 4(n) = Y(n) (9)
and for m = I;.., -1
N
Bn = iBn-1 + ~ n b n ] /[ l - ~nbnI (89)
f,( n ) k-1( n) - k,( n - 1) b,-1( n - 1) (92)
and
kn Pn + BnPn (90)
b,( n ) b,-,( n - I) - k,( n - I) fm-,( n). (93)
The reflectioncoefficients are updated tominimizethe
where K, and pn are the N-element and scalar partitions, sum of the mean-square value of the forward and backward
respectively, of the augmented Kalman gain vector defined prediction errors. That is,
in (SS).
Thematrixcomputations (82) involved in the Kalman k,( n ) = k,( n - 1)
algorithm are replaced in the fast Kalman algorithm by the + [ e-,(n)bm( n) + b:->( n - 1) L( n > ] / v m (94)
recursions (83) through (90) which use forward and back-
ward predictors t o update the Kalman gain vector as a new where
input sample y, is received and the oldest sample Y , , - ~ is D;n = w.,(n - 1 ) + If,-,(n)12 + Ib,-l( n - 1)12. (95)
discarded.
A fast exact initialization algorithm for the interval 0 d n The equalizer output for each stage is computed according
< N given in [IO]avoids the choice ofa stabilizing 6 in (79) to
and the resulting suboptimality of the solution at n = N. zm(n) = z,-1(n) + c,(n - l)brn(n)t
3) AdaptiveLattice Algorithms: The Kalman and fast
m = 0,1;-., N - 1. (96)
Kalman algorithms obtain their fast convergence by ortho-
gonalizing the adjustments made to the coefficients of an The final output ~ ~ - ~ is( usedn ) during data mode to
ordinary linear transversal equalizer. Adaptivelattice (AL) compute the receiver decision x(n). Initially, a reference
algorithms, on the other hand, use a lattice filter structure signal x(n) is substituted for the receiver decision in order
t o orthogonalize a set of received signal components [127]. t o compute the error signals
The vector b ( n ) is atransformedversionof the received K,( n) = wK,( n - 1) + tm(n - 1) fm-l( n ) (101)
vector y( n ) with elements y ( n ) , y ( n- I);. ., y(n - N + 1).
the forward prediction errors
This transformation is performed by an N X N lower trian-
gular matrix L according to b( n) = Ly( n), where 1 is formed fm( n) = fm-l( n) - G
,
( n - n - 1) (102)
bythebackwardpredictorcoefficientsoforder m, m = the backward prediction error
1 , . . ., N - 1 , The prediction error b,+,(n) is given by
m bm(n ) = n - 1) - H,( n - 1) f m - l ( n ) (103)
b,,,+,( n ) = Y( n - m ) - C B m ( j ) V( n - m + j ) the lattice coefficient for forward prediction
j-1
QURESHI: 1383
which have beenreportedinthe literature. In thetable tempting to state that no substantial further work remains
below, the number of operations (multiplications and divi- to be done. However, this has not been the case in the last
sions) required per iteration is listed for T- and T/Z-spaced decade despite how mature the field appeared in 1973 [59].
transversal equalizersof span NT seconds forthe LMS In fact, a number of the topics covered in this paper;e.g.,
gradient, Kalman, fast Kalman/fasttransversal, and RLS fractionally spacedequalizers, decision-aided IS1 cancella-
lattice algorithms. In each case, the smallest number of tion, and fast recursive leastsquares algorithms, were not
operations is given from the complexity estimates reported yetfullyunderstood or were yet to bediscovered. Of
in [78], [IO], and [52]. course, tremendous strides have since been made in imple-
mentationtechnologywhich have spawned newappli-
Number of Operations per Iteration
Algorithm Equalizer T/2 T Equalizer cations, e.g., digital subscriber loops, and pushed existing
applicationstowardtheirlimits, e.g., 256-QAMdigital ra-
L M S gradient 2N 4N
Kalman 2N2 + 5N 8N2 + 10N dios and voice-band modems with rates approaching 19.2
Fast Kalman/fast transversal 7 N + 14 24N + 45 kbits/s. Programmable digital signal processors now permit
RLS lattice 15N - 11 46N implementation of ever more sophisticated and computa-
tionally complex algorithms; and so the study and research
The fast Kalman is themostefficienttype of RLS al- must continue-in new directions. There is still more work
gorithm.However, compared to the LMS gradient al- to be done in adaptive equalization of nonlinearities with
gorithm, the fast Kalman algorithm is still about four times memory and in equalizer algorithms for coded modulation
as complex for Tequalizers and six times for T/2 equalizers. systems. However, the emphasis has already shifted from
The RLS latticealgorithm is still in contention due to its adaptive equalization theory toward the more general the-
better numerical stability and order-recursive structure, de- ory and applications of adaptive filters, and toward struc-
spite a two-fold increase in computational complexity over tures and implementation technologies which are uniquely
the fast Kalman algorithm. suited to particular applications.
The discussion on RLS algorithms would not be complete
without a comment on the numerical problems associated
with these algorithms in steady-state operation. Simulation ACKNOWLEDGMENT
studies have reported the tendency of RLS algorithms im-
plemented with finite precision to becomeunstableand Theauthorwishes to thank C. D. Forney, Jr., forhis
the adaptive filter coefficients to diverge [IO], [&I, [52], [78], encouragement and
guidance in preparing this
paper.
[126].This is due to the long-term accumulation of finite Thanks are also due to a number of friends and colleagues
precision errors. Amongthedifferent types of RLS al- who have reviewed and commented on the paper.
gorithms, the fast Kalman or fast transversal type algorithms
are the most prone to instability [&I, [78]. In [78] instability REFERENCESAND BIBLIOGRAPHY
was reportedto occur when an exponential weighting
factor w 1 wasused for the fast Kalman algorithm imple- K.Abendand 6. D. Fritchman,“Statisticaldetectionfor
mented with single-precision floating-point arithmetic. The communicationchannels with intersymbolinterference,”
P ~ O C .IEEE, VOI. sa, pp. 779-785, May 1970.
Kalman and RLS lattice algorithms did not show this insta- 0. Agazzi, D. A.Hodges,and D. G.Messerechmitt,”Large
bility. A sequential processing dual-channel version of the scale integration of hybrid-method digitalsubscriber loops,”
RLS lattice algorithm for a DFE is reported to be stable even /E€€ Trans. Commun., vol. COM-30,pp.2095-2108,Sept.
for fixed-point arithmetic with I O - to 12-bit accuracy [52]. 1982.
M. E. Austin,“Decision-feedbackequalizationfordigital
However, in [IO] an “unnormalized” RLS lattice algorithm is
communication over dispersive channels,” MIT Lincoln Lab.,
shownto becomeunstable.Normalized versions of fast Lexington, MA, Tech. Rep. 437, Aug. 1967.
transversal [IO] and RLS lattice[I271algorithms are more C. A. Belfiore and J. H. Park, Jr., “Decision feedback equali-
stable but both require square roots, the lattice type having zation,” Proc. /€E€, vol. 67, pp. 1143-1156, Aug. 1979.
greater computational complexity. E. Biglieri, A. Cersho, R. D. Citlin, andT.L.Lim,“Adaptive
cancellation nonlinear
of intersymbol
interference
for
The stability of RLS algorithms can be improved [IO], [ a ] , voiceband data transmission,” /E€€ 1. SelectedAreasCom-
[I261 bymodificationofthe leastsquares criterion,and mun., voi. SAC-2, pp. 765-777, Sept. 1984.
periodic reinitialization of the algorithm to avoid precision D. M. Brady, “Anadaptivecoherentdiversityreceiverfor
error buildup. The modified criterion takes into account the datatransmissionthroughdispersivemedia,” in Proc. 7970
/€E€ Int. Conf. Commun., pp. 21-35 to 21-39, June 1970.
squaredmagnitudeofthedifferenceofthefiltercoeffi-
C. Caraiscos and E. Liu, “A roundoff error analysis of the L M S
cients from their initial (or restart) values. The rationale for adaptive algorithm,” /€E€ Trans. Acoust., Speech, Signal Pro-
thisso-called soft constraint[IO] is the same as forthe cess., vol. ASSP-32, pp. 34-41,Feb.1984.
stochastic gradient algorithm with tap leakage discussed in R. W. Chang,“Anewequalizerstructureforfaststart-up
Section Ill-C. digital communication,” Bell Syst. Tech. I., vol. 50, pp.
1%9-2014, JUly-AUg. 1971.
For a short transition period following each reinitializa- R. W. Chang and E. Y . Ho, “On fast start-up data communi-
tion, while the RLS algorithm is reconverging, an auxiliary cationsystemsusingpseudo-randomtrainingsequences,”
L M S adaptive filter is used to compute outputs. Results of Bell Sysr. Tech. I., vol. 51, pp. 2013-2027, Nov. 1972.
four variations oftheperiodic restart procedurefor an J. M. Cioffi and T. Kailath,“Fast,recursive-least-squares,
adaptive decision feedback equalizer are given in [126]. transversal filters for adaptive filtering,” /E€€ Trans. Acoust.,
Speech, Signal Process., vol. ASSP-32, pp. 304-337, Apr. 1984.
J. M. Cioffi and T. Kailath, ”An efficient exact-least-squares
fractionallyspacedequalizerusingintersymbolinterpola-
V. CONCLUDINGREMARKS tion,“ /€E€ J, SelectedAreas Commun., vol. SAC-2, pp.
743-756, Sept. 1984.
Adaptiveequalizationandthemore general fieldof D. L. Duttweiler, J. E. Mazo, and D. C. Messerschmitt, “An
adaptive filtering have been areas of active research and upper bound on the error probability in decision-feedback
development for more than two decades. It is, therefore, equalization,” /€€E Trans.Inform.Theory, vol. IT-20, pp.