17
Equalization in Communication Engineering

Deconvolution (Section 5.2.3 and Section 10.1) is the linear filter that compensates for a waveform distortion of convolution. In communication systems, equalization refers to the compensation of the convolutive mixing due to the signals’ propagation (in this context it is called the communication channel), either in wireless or wired digital communications. Obviously, deconvolution is the same as equalization, even if it is somewhat more general as being independent of the specific communication system setup. In MIMO systems (Section 5.3) there are multiple simultaneous signals that are propagating from dislocated sources and cross‐interfering; the equalization should separate them (source separation) and possibly compensate for the temporal/spatial‐channel convolution. The focus of this chapter is to tailor the basic estimation methods developed so far to the communication systems where single or multiple channels are modeled as time‐varying, possibly random with some degree of correlation. Signals are drawn from a finite alphabet of messages and are thus non‐Gaussian by nature [57] . To limit the contribution herein to equalization methods, the estimation of the communication channel (or channel identification, see Section 5.2.3) is not considered as being estimation in a linear model with a known excitation. In common communication systems, channel estimation is part of the alternate communication of a training signal interleaved with the signal of interest as described in Section 15.2. There are blind‐estimation methods that can estimate the channel without a deterministic knowledge of transmitted signals, relying only on their non‐Gaussian nature. However, blind methods have been proved to be too slow in convergence to cope with time‐variation in real systems and thus of low practical interest for most of the communication engineering community.

17.1 Linear Equalization

In linear modulation, the transmitted messages are generated by scaling the same waveform for the corresponding information message that belongs to a finite set of values (or alphabet) images with cardinality images that depends on symbol mapping (Section 15.2). It is common to consider a multi‐level modulation with images, or images if complex valued. The transmitted signal is received after convolution by the channel and symbol‐spaced sampling that is modeled as (see Section 15.2):

images

with images and white Gaussian noise

images

Linear equalization (Figure 17.1) needs to compensate for the channel distortion by an appropriate linear filtering images as

images

so that the equalization error

images

is small enough, and the equalized signal coincides with the undistorted transmitted one up to a certain degree of accuracy granted by the equalizer.

Communication system model displaying 2 boxes labeled channel H(z) and equalization G(z) with a crossed circle in between, and with connecting arrows labeled a[n], w[n], x[n], and aˆ[n].

Figure 17.1 Communication system model: channel H(z) and equalization G(z).

Inspection of the error highlights two terms:

The first is the noise filtered by the equalizer g[n], and the second one is the incomplete cancellation of the channel distortions. Design of linear equalizers as linear estimators of a[n] have been already discussed as deconvolution (Section 12.2); below is a review using the communication engineering jargon. In digital communications, the symbols a[n] are modeled as a stochastic WSS process that is multivalued and non‐Gaussian; samples are zero‐mean and independent and identically distributed (iid):

images

and the autocorrelation sequence is

images

There are two main criteria adopted for equalization based on which the error of (17.1) is accounted for in optimization.

17.1.1 Zero Forcing (ZF) Equalizer

The zero‐forcing (ZF) equalizer is

images

called this way as it nullifies any residual channel distortion: images. The benefit is the simplicity but the drawbacks are the uncontrolled amplification of the noise that is filtered by the equalizer with error images (e.g., any close to zero H(ω) enhances uncontrollably the noise as images), and the complexity in filter design that could contain causal/anti‐causal responses over an infinite range.

17.1.2 Minimum Mean Square Error (MMSE) Equalizer

The minimization of the overall error images accounts for both noise and residual channel distortion. The z‐transform of the autocorrelation of the error:

images

depends on filter response {g[n]}. The minimization of images wrt the equalizer yields to the orthogonality condition as for the Wiener filter (Section 12.2):

images

and the (linear) MMSE equalizer is

images

The MMSE equalizer converges to

images

for the two extreme situations of small or large noise, respectively and images is called matched filter as it acts as a filter that correlates exactly with H(z). Needless to say, linear MMSE is sub‐optimal and some degree of non‐linearity that accounts for the alphabet images is necessary to attain the performance of a true MMSE equalizer.

17.1.3 Finite‐Length/Finite‐Block Equalizer

This is the convolution of a finite block of symbols a with the finite length channel h, and its equalization is

images

The equalization using the ZF condition seeks for

images

and the MMSE

images

To solve these minimizations, one has to replace the notation with the convolution matrix for the channel h as in Section 5.2 and Section 5.2.3. The optimizations reduce to the one for MIMO systems and solutions are detailed below in Section 17.3.

17.2 Non‐Linear Equalization

Given the statistical properties of signals, the optimal Bayesian estimator is non‐linear and resembles the one derived in Section 11.1.3. However, some clever reasoning can simplify the complexity of non‐linear Bayesian estimator by the so called Decision Feedback Equalization (DFE). The channel response images contains causal (post‐cursor) and anti‐causal (pre‐cursor) terms, and the DFE exploits the a‐priori information on the finite alphabet images for some temporary linear estimates based on the causal part of h[n]. The approximation of the Bayesian estimator is to use the past to predict the current degree of self‐interference for its reduction, after these temporary estimates are mapped onto the finite alphabet images.

Referring to the block diagram in Figure 17.2, the equalizer is based onto two filters C(z) and D(z), and one mapper onto the alphabet images (decision). The decision is a non‐linear transformation that maps the residual onto the nearest‐level of the alphabet (see Figure 17.3 for images). The filter C(z) is designed based on the ZF or MMSE criteria, and the filter images is strictly causal as it uses all the past samples in attempt to cancel the contribution of the past onto the current sample value and leave ɛ[n] for decisions. When the decision â[n] is correct, the cancellation reduces the effect of the tails of the channel response but when noise is too large, the decisions could be wrong and these errors accumulate; the estimation is worse than linear equalizers. This is not surprising as the detector replaces a Bayes estimator with a sharp piecewise non‐linearity in place of a smooth one, even when the noise is large (see Section 11.1.3). Overall, the DFE reduces to the design of two filters.

Flow diagram of decision feedback equalization displaying 5 boxes labeled H(z), T(z), 1 +D(z), and D(z), 2 crossed circles, and connecting arrows labeled a[n], w[n], x[n], f [n], etc.

Figure 17.2 Decision feedback equalization.

Cartesian plane of ε vs. â of decision levels for A = {±1,±3} displaying a stair-step line along the diagonal dashed line.

Figure 17.3 Decision levels for images.

The filter C(z) is decoupled into two terms

images

where images acts as a compensation of the feedback filter D(z), and the filters to be designed are T(z) and D(z). Assuming that all decisions are correct images (almost guaranteed for noise images small enough), the z‐transform of the error

contains two terms: the residual of the equalization, and the filtered noise. The design depends on how to account for these two terms.

17.2.1 ZF‐DFE

The filter T(z) can be designed to nullify the residual of the filter equalization (zero‐forcing condition); then the design of the causal filter D(z) is a free parameter to minimize the filtered noise. More specifically, the choice is

images

and the filter D(z) should be chosen to minimize the power of the filtered noise images with power spectral density in z‐transform

images

The structure of the filter images is monic and causal, it can be considered as a linear predictor (Section 12.3) for the process f[n] correlated by T(z). The factorization images into min/max phase terms according to the Paley–Wiener theorem (Section 4.4.3) (recall that images but it is not necessarily true that images, see Appendix of Chapter 4) yields to the choice

images

Alternatively, and without any change of the estimator, one can make the factorization of the filter into min/max phase images and the filter becomes

images

Filters are not constrained in length, but one can design images using a limited‐length linear predictor at the price of a small performance degradation.

17.2.2 MMSE–DFE

In the MMSE method, the minimization is over the ensemble of error (17.2). Namely, the error can be rewritten by decoupling the terms due to the filter T(z)

images

and the linear predictor images as

images

The power spectral density of f[n] depends on t[n]:

images

and the MMSE solution is from orthogonality, or equivalently the minimization

images

where images, and images.

The linear predictor images follows by replacing the terms. The (minimal) PSD for the filter images is

images

and its min/max phase factorization defines the predictor

images

The filter C(z) for the MMSE–DFE equalizer is

images

which is a filter with causal/anti‐causal component.

Remark. Since the filter’s length is unbounded and can have long tails, the implementation of the MMSE–DFE could be quite expensive in terms of computations. A rule of thumb is to smoothly truncate the response cMMSE[n] by a window over an interval that is in the order of 4–6 times the length of the channel h[n]. In other cases, the design could account for the limited filter length with better performance compared to a plain truncation, even if smooth.

17.2.3 Finite‐Length MMSE–DFE

The MMSE–DFE can be designed by constraining the length of the filters to be limited [75] . Making reference to Figure 17.4, the images filter c can have a causal and anti‐causal component so that

images

where the last equality is for correct decisions, and images. The error

images

is represented in terms of the augmented terms images and images, and the MSE becomes

images
Flow diagram of finite length MMSE–DFE displaying 4 boxes labeled h↔H(z), c↔C(z), and d↔D(z) with 3 double-headed arrows labeled Nh, Nc, and Nd, and other connecting arrows.

Figure 17.4 Finite length MMSE–DFE.

The optimization wrt c for known d (the constraint condition is usually added at last step) is quadratic:

images

and the MSE

is a quadratic form, but the constraint condition images has not been used yet.

Rewriting explicitly the MSE (17.3)

images

after partitioning

images

it yields the solution

images

and the MSE for this optimized solution becomes images.

In the case that the channel has length Nh samples, the empirical rules are

images

with images. However, since the optimal MSE has a closed form, the filters’ length can be optimized for the best (minimum) MSEmin within a predefined set; this is the preferred solution whenever this is practicable.

17.2.4 Asymptotic Performance for Infinite‐Length Equalizers

The performance is evaluated in terms of MSE. For a linear MMSE equalizer, the MSE is

images

where images, and by using the solutions derived above (argument z is omitted in z‐transforms)

images

it is

images

The steps for the MMSE–DEF equalizer are the same as above by replacing the solutions for the images and after some algebra [74] :

images

Since

images

according to the Jensen inequality for convex transformation exp(.) (Section 3.2) for images this proves that

images

The DFE should always be preferred to the linear MMSE equalizer except when noise is too large to accumulate errors.

17.3 MIMO Linear Equalization

MIMO refers to linear systems with multiple input/output as discussed in model definition in Section 5.3, a block of N symbols a are filtered by an images matrix H with images (i.e., more receivers that transmitted symbols) and received signals are

with noise images. MIMO equalization (Figure 17.5) refers to the linear or non‐linear procedures to estimate a from x provided that H is known. In some contexts the interference reduction of the equalization is referred as MIMO decoding if the procedures are applied to x, and MIMO precoding if applied to a before being transmitted as transformation g(a) with model images, possibly linear images. The processing variants for MIMO systems are very broad and are challenging many researchers in seeking different degrees of optimality. The interested reader might start from [72,73].

Flow diagram of Linear MIMO equalization displaying four vertical blocks labeled from a (left) with an intersecting arrows to x (middle), to G, and to â (right).

Figure 17.5 Linear MIMO equalization.

To complete the overview, the model (17.4) describes the block (or packet‐wise) equalization when a block of N symbols a are arranged in a packet filtered with a channel of images samples, and the overall range is images; the matrix H in (17.4) is the convolution matrix in block processing.

17.3.1 ZF MIMO Equalization

Block linear equalization is based on a transformation images such that the ZF condition holds: images. The estimate is

images

and the metric to be optimized is the norm of the filtered noise images. The optimization is constrained as

images

to be solved by the Lagrange multiplier method by augmenting the metric with N2 constraints using the images matrix images :

images

Optimization wrt G is (Section 1.8 and Section 1.8.2)

images

using the constraint condition

images

it yields the ZF equalization

images

Note that the ZF solution can be revised as follows. Since images, the Cholesky factorization images into lower/upper triangular matrixes is unique, and similarly images. The ZF estimate is

images

where images and images; it can be interpreted as the pseudoinverse of images for the pre‐whitened observation images.

17.3.2 MMSE MIMO Equalization

The linear MMSE estimator has already been derived in many contexts. To ease another interpretation, the estimator is the minimization of the MSE

images

for the symbols with correlation images. The MMSE solution follows from the orthogonality (Section 11.2.2)

images

This degenerates into the ZF for large signals (or images) as images; and for large noise it becomes images. The MSE for âMMSE can be evaluated after the substitutions.

17.4 MIMO–DFE Equalization

The MIMO–DFE in Figure 17.6 is just another way to implement the cancellation of interference embedded into MIMO equalization, except that the DFE cancels the interference from decisions—possibly the correct ones. In MIMO systems this is obtained from the ordered cancellation of samples, where the ordering is a degree of freedom that does not show up in equalization as in these systems, samples are naturally ordered by time. Before considering the DFE, it is essential to gain insight into analytical tools to establish the equivalence between DFE equalization and its MIMO counterpart.

Flow diagram of MIMO–DEF equalization from a rightward arrow labeled a to a box (H) passing through a crossed circle to a dotted box of C=LT, another crossed circle, and to a box labeled B (left to right).

Figure 17.6 MIMO–DEF equalization.

17.4.1 Cholesky Factorization and Min/Max Phase Decomposition

The key algebraic tool for MIMO–DFE is the Cholesky factorization that transforms any images into the product of two triangular matrixes, one the Hermitian transpose of the other—sometimes referred as the (unique) square‐root of R:

images

where the entries of the diagonal matrix images are all positive. For zero‐mean rvs, Cholesky factorization of the correlation matrix is the counterpart of min/max phase decomposition of the autocorrelation sequence. To prove this, let images be an images vector with covariance images. Transformation by a lower triangular matrix L with normalized terms along the diagonal is

images

and the covariance is

images

If it is now assumed that a vector y has covariance images, the transformation can be inverted as images —still lower triangular (images for images). The i th entry is

images

where images extracts the i th row from images. Since images and images then images is a linear predictor that uses all the samples images (similar to a causal predictor over a limited set of samples) to estimate yi, and images is the variance of the prediction error. If choosing images the linear predictor makes use of the complementary samples images to estimate yi and the predictor would be reversed (anti‐causal). For a WSS random process, the autocorrelation can be factorized into the convolution of min/max phase sequences, and these coincide with the Cholesky factorization of the Toeplitz structured correlation matrix R for images. However when N is small, the Cholesky factorization of the correlation matrix R takes the boundary effects into account, and the predictors (i.e., the rows of images) are not just the same shifted copy for every entry.

17.4.2 MIMO–DFE

In MIMO–DFE, the received signal x is filtered by a compound filter to yield

images

where the normalized (images for images) lower triangular matrix L is decoupled into the strictly lower triangular B and the identity I to resemble the same structure as DFE in Section 17.2. The matrix DFE acts sequentially on decisions as the product

images

isolates the contributions and sequentially orders the contributions arising from the upper lines toward the current one: images. The fundamental equation of DFE is the equality of the decision variable from the previous decisions carried out from the upper lines:

images

From the partition above:

images

according to the model

images

There are two metrics to be optimized based on the double constraint on the structure of L according to the ZF or MMSE constraint:

images

For both criteria the second step, which is based on the first solution, is peculiar for DFE and it implies the optimization of a metric with the constraint on the normalized lower triangular structure of the matrix L. Namely, in both cases it follows the minimization

for a certain positive definite matrix images that is, for the two cases and images:

images

Since images, its Cholesky factorization, is

images

where images with images. On choosing

images

it follows that

images

Solutions for AWGN images are derived below.

ZF–DFE for AWGN: images

The optimization

images

yields the same solution as in Section 17.3.1:

images

Since the noise after filtering TZF is images, the metric of (17.5) becomes images and the filter LZF follows from its Cholesky factorization

images

MMSE–DFE for AWGN: images

The solution of the first term is straightforward from orthogonality (or the MMSE–MIMO solution):

images

using here any of the two equivalent representations (Section 11.2.2), with images. The matrix images, where the second equality follows from the matrix inversion lemma (Section 1.1.1), and LMMSE is from the Cholesky factorization.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
3.129.19.251