Chapter 4

Recent Advances on LIP Nonlinear Filters and Their Applications

Efficient Solutions and Significance-Aware Filtering

Christian HofmannWalter Kellermann    Friedrich-Alexander-Universität Erlangen-Nürnberg, Lehrstuhl für Multimediakommunikation und Signalverarbeitung, Erlangen, Germany

Abstract

Linear-In-the-Parameters (LIP) nonlinear filters are categorized as Cascade Models (CMs) (generalizing Hammerstein models), Cascade Group Models (CGMs) (generalizing Hammerstein Group models (HGMs) and including, e.g., Volterra filters) and bilinear cascade models, where the filter output is a bilinear function of the model parameters. Time-domain and partitioned-block frequency-domain adaptation of CGMs and CMs is described and the methods for adapting bilinear cascade models are summarized as variants of the filtered-X adaptation. These models and algorithms are employed to review the Significance-Aware (SA) filtering concept, decomposing the model for the unknown system and the adaptation mechanism into synergetic subsystems to achieve high computational efficiency. In particular, the Serial SA (SSA) and Parallel SA (PSA) decomposition lead to SSA-CMs, PSA-CGMs and a novel PSA filtered-X algorithm. The main concepts described in this chapter are exemplarily compared for the challenging application of nonlinear acoustic echo cancellation. Furthermore, model structure estimation for LIP nonlinear filters based on convex filter combinations is briefly outlined and compared with SA filtering.

Keywords

Cascade models; Cascade group models; Hammerstein models; Hammerstein group models; Significance-aware filtering; SA filtering; Filtered-X

Chapter Points

  • •  LIP nonlinear filters with simple linear dependence on the parameters can be identified in a unified way by classical partitioned-block frequency-domain normalized least mean square algorithms.
  • •  Algorithms to adapt LIP nonlinear filters with bilinear dependence on parameter subsets can be treated in a unified way as filtered-X algorithms.
  • •  Significance-aware filtering can be applied in both cases for the efficient and effective identification of LIP nonlinear filters.

Acknowledgements

This book chapter could not have been written without the dedicated work of many former students and collaborators in the authors' research group, especially Alexander Stenger, Fabian Küch, Marcus Zeller and Luis A. Azpicueta-Ruiz.

4.1 Introduction

Linear-In-the-Parameters (LIP) nonlinear filters constitute a broad class of nonlinear systems and are used for modeling in a broad range of scientific areas, such as in biology (e.g., [1,2]), control engineering (e.g., [3]), communications (e.g., [46]) and in acoustic signal processing, where they are, for instance, employed for loudspeaker linearization (e.g., [79]) and Acoustic Echo Cancellation (AEC) (e.g., [1016]). A nonlinear system with input x(n)Image and output y(n)Image, where n is the discrete-time sample index, will be described by the input/output relation

y ( n ) = f I/O { x ( n ) } ,

Image (4.1)

and will be called a memoryless nonlinearity if and only if y(n)δ(n)=fI/O{x(n)δ(n)}Image for any n and x(n)Image, where δ(n)Image is the unit impulse. Otherwise, fI/O{x(n)}Image will be called a nonlinearity with memory. In both cases, a nonlinearity may either be considered as time-invariant and known, or as parametrized by a set of unknown parameters. The following treatment will focus on LIP nonlinear filters, where the output depends linearly on the parameters, e.g., on a vector a, or bilinearly on two parameter subsets, e.g., on parameter vectors a1Image and a2Image.

The adaptive identification of a parametric model for an unknown nonlinear system fI/O{}Image is illustrated in Fig. 4.1. Therein, an adaptive nonlinear filter (the parametric model) produces an estimate yˆ(n)Image for an observation y(n)Image and the filter's parameters are iteratively refined (adapted) to minimize a cost function derived from the error signal e(n)=y(n)yˆ(n)Image, based on noisy observations y(n)=fI/O{x(n)}+s(n)Image. When obtaining the observations by measuring a physical signal, s(n)Image may be the superposition of sensor self-noise and other processes coupling into the sensors as well.

Image
Figure 4.1 The generic task of adaptive system identification.

As will be explained in the following, AEC is a particularly challenging example for system identification. For AEC, y(n)Image in Fig. 4.1 corresponds to a microphone signal composed of local sound sources s(n)Image (e.g., the local speaker in a telephone communication) and acoustic echoes fI/O{x(n)}Image from a played-back loudspeaker signal x(n)Image (e.g., the far-end speaker in a telephone communication). The AEC system aims at removing the acoustic echoes fI/O{x(n)}Image from the microphone signal y(n)=fI/O{x(n)}+s(n)Image, using an echo signal estimate yˆ(n)Image. In periods of local sound source activity, the error signal e(n)Image constitutes an estimate for the local sound sources s(n)Image. In single-talk periods where only the echo signal fI/O{x(n)}Image contributes to the microphone signal, the error signal can be employed to refine the adaptive filter. The detection of local sound sources (double-talk detection) is out of the scope of this chapter and solutions for this problem exist (e.g., [17,18]). Therefore, we will assume that no local sources are active when considering AEC, which means s(n)=0Image.

AEC is a particularly challenging example for system identification due to the complex acoustic echo paths to be modeled, which typically requires nonlinear adaptive filters with large memory. Furthermore, adaptive filters have to identify a system excited by speech or audio signals (super-Gaussian, non-white and highly instationary signals) instead of signals tailored to system identification. Thus, constituting an especially challenging application, AEC will be used as a reference for evaluating the methods in this chapter. However, the models described and introduced in this chapter are not limited to AEC. Potential other applications will be highlighted at the beginning of Section 4.5, where the required terminology on nonlinear systems and their adaptation will have been established.

The organization of this chapter will be summarized in the following. After introducing the fundamental notations and terminology in the remainder of this section, state-of-the-art LIP nonlinear filters will be categorized in Section 4.2 and fundamental algorithms for parameter estimation for such filters will be presented in Section 4.3. Resulting from a unified treatment of time-domain Normalized Least Mean Square (NLMS), Partitioned-Block Frequency-domain Normalized Least Mean Square (PB-FNLMS), as well as Filtered-X (FX) algorithms for LIP nonlinear filters, a novel and efficient FX PB-FNLMS algorithm will be introduced. In Section 4.4, the recently proposed, computationally very efficient concept of Significance-Aware (SA) filtering will be summarized and a novel SA FX PB-FNLMS algorithm with even further increased efficiency will be proposed. The adaptive filtering techniques described in this chapter will be evaluated and compared in Section 4.5. Finally, similarities and differences of SA filters and nonlinear adaptive filtering techniques based on convex filter combinations for model structure estimation will be reviewed in Section 4.6, before summarizing this chapter in Section 4.7.

Notations and Fundamentals  As common in system identification literature, the terms filter and system are used interchangeably. The input/output relationship of a linear, causal Finite Impulse Response (FIR) system of length L with input signal x(n)Image and output y(n)Image will be written as

y ( n ) = h ( n ) x ( n ) = k = 0 L 1 h ( k ) x ( n k ) = h , x ˇ n ,

Image (4.2)

where h(k)Image denotes the kthImage tap of the system's Impulse Response (IR), ⁎ denotes convolution, and h,xˇnImage represents the scalar product between the vectors h and xˇnImage, where

h = [ h ( 0 ) , h ( 1 ) , , h ( L 1 ) ] T ,

Image (4.3)

x ˇ n = [ x ( n ) , x ( n 1 ) , , x ( n L + 1 ) ] T

Image (4.4)

denote the IR vector and the time-reversed input signal vector, respectively, and the time reversal is indicated by the caron accent. Furthermore, 0MImage and IMImage will denote the M×MImage matrix of zeros and the M×MImage identity matrix, respectively, and are used as components of the windowing matrices

W 01 = [ 0 M 0 M 0 M I M ] and W 10 = [ I M 0 M 0 M 0 M ] ,

Image (4.5)

which set the first and second half of a vector of length N=2MImage to zero, respectively. In addition, F will refer to the N×NImage unitary Discrete Fourier Transform (DFT) matrix and a1a2Image and a1a2Image will denote element-wise multiplication and division of the vectors a1Image and a2Image, respectively. The element-wise magnitude square computation for the entries of a vector a will be written as |a|2Image.

4.2 A Concise Categorization of State-of-the-Art LIP Nonlinear Filters

In this section, generic state-of-the-art LIP nonlinear filter structures are introduced as building blocks for the recently developed and newly proposed adaptive filtering methods in Sections 4.3 and 4.4.

4.2.1 Hammerstein Models and Cascade Models

A very simple class of LIP nonlinear filters can be described with the cascade structure depicted in Fig. 4.2, where a nonlinearity f{}Image is followed by a linear system with IR vector h. The output depends linearly on the parameters h of the linear system and can be expressed in the notation of Eq. (4.2) as

y ( n ) = f { x ( n ) } h ( n ) = x NL ( n ) h ( n ) .

Image (4.6)

If f{}Image in Fig. 4.2 is a memoryless nonlinearity, the structure is widely known as HM [19], whereas the more general case, where f{}Image is a nonlinearity with memory, will simply be referred to as CM in the following (the attribute nonlinear will be implied). HMs can, for instance, be employed for loudspeaker linearization (e.g., [20,21]) and AEC (e.g., [2232]). HMs are simple and powerful models for acoustic echo paths if the major source of nonlinearity is the playback equipment (amplifiers and loudspeakers operating at the limits of their dynamic range) and can be mapped to f{}Image, while the acoustic IR (describing sound propagation through the acoustic environment) can be described by h [22].

Image
Figure 4.2 A CM, which is commonly referred to as HM for the special case of a memoryless nonlinearity f{⋅}.

4.2.2 Hammerstein Group Models and Cascade Group Models

The term Cascade Group Model (CGM) will refer to models with the structure depicted in Fig. 4.3. Such CGMs are composed of B parallel branches, each of which is a nonlinear CM (see Section 4.2.1) with a branch nonlinearity fb{}Image and IR h(b)Image. For memoryless branch nonlinearities fb{}Image, each branch becomes an HM and the entire CGM will be referred to as HGM in the following. HGMs have successfully been employed to describe both acoustic systems, e.g., in [3337], and nonacoustic systems, e.g., [38,39]. In their treatment in the literature, HGMs have been addressed by various names, such as the Uryson model [3,39], or are simply treated as an approximation of a particular CGM. The most prominent examples of such HGM/CGM pairs are the so-called power filters [37,40,41] and Volterra filters [19]. Power filters are HGMs with monomes as branch nonlinearities fb{}Image, e.g.,

f 1 { } = ( ) , f 2 { } = ( ) 2 , f 3 { } = ( ) 3 , , . f B { } = ( ) B .

Image

Image
Figure 4.3 The general structure of a CGM, which is also referred to as HGM for the special case of memoryless nonlinearities fb{⋅},b ∈ {1,…,B}.

Volterra filters result from power filters by augmenting the latter with additional branches where the branch nonlinearities fb{}Image perform time-lagged products of powers of the input (time-lagged products of the power filter branch nonlinearities), e.g., the set

f 1 { x ( n ) } = x ( n ) , f 2 { x ( n ) } = x 2 ( n ) , f 3 { x ( n ) } = x ( n ) x ( n 1 ) , , f 2 + R { x ( n ) } = x ( n ) x ( n R ) , f 3 + R { x ( n ) } = x 3 ( n ) ,

Image

leads to a Volterra filter in Diagonal Coordinate Representation (DCR) [6]. The set of IRs h(b)Image of branches where fb{}Image performs (potentially time-lagged) products of o input signal values is called oth-order Volterra kernel and the branches within a kernel are also termed diagonals of the respective kernel. In the aforementioned example, the second-order kernel is modeled by R+1Image diagonals (b{2,,2+R}Image). Volterra filters are universal approximators for a large class of nonlinear systems and have therefore been applied in biology (e.g. [1]), control engineering (e.g., [3]) and communications (e.g., [46]), as well as in acoustic signal processing (e.g., [716,42,43]). Recently, HGMs and CGMs have also been realized with other sets of branch nonlinearities, such as Fourier basis functions [44], Legendre polynomials [45,46] or Chebyshev polynomials [47]. Also, Functional-Link Adaptive Filters (FLAFs) [34,35,4850] can be viewed as HGM or CGM structures. All these models can be described by the input/output relationship

y ( n ) = b = 1 B k = 0 L 1 f b { x ( n k ) } x b ( n k ) h ( b ) ( k ) = b = 1 B x b ( n ) h ( b ) ( n ) ,

Image (4.7)

where h(b)(k)Image denotes the kth tap of the linear subsystem in branch b and xb(n)Image is the bth branch signal. Thus, algorithms developed for one particular realization (i.e., set of basis functions fb{}Image) of CGMs can often immediately be applied to other realizations. This holds for the adaptation algorithms for coefficient estimation described in Sections 4.3 and 4.4, as well as for the structure estimation methods outlined in Section 4.6.

4.2.3 Bilinear Cascade Models

For applications where CGMs with a large number of branches are computationally too demanding, e.g., for AEC in mobile devices with very limited processing or battery power, CM-structured adaptive filters are particularly attractive. Yet, the nonlinearity itself is usually unknown. Therefore, the nonlinearity f{}Image of a CM is typically modeled by a parametric preprocessor according to

f { x ( n ) } = f pp { x ( n ) } = b = 1 B w b f b { x ( n ) } = x pp ( n ) ,

Image (4.8)

where B nonlinearities fb{}Image are combined with weights wbImage, which are parameters to be identified in addition to the linear subsystem. The subindex “pp” of the nonlinear preprocessor and its output signal are used to emphasize the parametrical structure of the preprocessor. The general structure of a CM with a preprocessor according to Eq. (4.8) is depicted in Fig. 4.4A. Most prominently, fb{}Image are chosen as monomes [20,21,24,5154], e.g., fb{}=()bImage, such that the nonlinearity fpp{}Image can be seen as a truncated Taylor series expansion. Fourier series- [55] and Legendre series-based approximations [28,30,31,56] have also recently been employed in AEC. Expanding the preprocessor of Fig. 4.4A in the block diagram results in Fig. 4.4B. Obviously, such a system has the input/output relationship

y ( n ) = x pp ( n ) h ( n )

Image (4.9)

= h ( n ) b = 1 B w b f b { x ( n ) } x b ( n ) y b ( n ) ,

Image (4.10)

and the paths from xb(n)Image to y(n)Image consist not only of a single linear system, but of cascades of wbImage (a single-tap linear system) with the linear system h. Therefore, the output of y(n)Image linearly depends on both the parameter vector w=[w1,w2,,wB]TImage and the parameter vector h, which is why such systems will be referred to as bilinear1 CMs in the following. Bilinearity holds regardless of fb{}Image being memoryless or with memory and also for general linear filters with IRs h(b)Image replacing the single-tap filters wbImage, resulting in CGM preprocessors.2 Thus, in the following, all considerations on simple preprocessor systems according to Fig. 4.4A translate to bilinear CMs with CGM preprocessors and vice versa, unless noted otherwise.

Image
Figure 4.4 Alternative representations of a CM with parametric preprocessor. (A) Compact representation of a CM with parametric preprocessor as nonlinear–linear cascade. To emphasize the parametric structure of the preprocessor, the nonlinearly distorted input signal is denoted xpp(n)Image instead of xNL(n)Image. (B) Expansion of the memoryless nonlinearity in (A), revealing a bilinear dependency of the model output y(n)Image on the parameter vectors w=[w1,w2,,wB]TImage and h.

4.3 Fundamental Methods for Coefficient Adaptation

The most common methods for parameter estimation for linear and LIP nonlinear filters are stochastic gradient-type algorithms, such as the Least Mean Square (LMS) algorithm or the NLMS algorithm [57]. These can be derived as an approximation of a gradient descent w.r.t. the mean squared error cost function J=E{|e(n)|2}Image, where E{}Image denotes mathematical expectation and e(n)=y(n)yˆ(n)Image is the error signal, computed as difference between the unknown system's output y(n)Image and its estimate yˆ(n)Image, obtained with the adaptive filter (cf. Fig. 4.1). For a comprehensive treatment of different types of LMS-type algorithms, we refer to [57].

In the following, classical system identification will be employed in Section 4.3.1 to adapt h(b)Image of CGMs from Section 4.2.2 (cf. Fig. 4.3). In Section 4.3.2, the filtered-X method is employed to convert the adaptation of a preprocessor system according to Section 4.2.3 (cf. Fig. 4.4) into two ordinary system identification tasks, one for h(b)Image and one for wbImage, by reordering wbImage and h(b)Image.

4.3.1 NLMS for Cascade Group Models

In this section, the adaptation of CGMs according to Section 4.2.2 by a time-domain and a partitioned-block frequency-domain NLMS will be described in Sections 4.3.1.1 and 4.3.1.2, respectively. Note that the branch nonlinearities will be considered as time-invariant and known, such that only the linear subsystems of h(b)Image of the CGMs from Section 4.2.2 are adaptive.

4.3.1.1 Time-Domain NLMS

To implement and adapt a CGM, the input signal x(n)Image has to be mapped nonlinearly to a set of branch signals

x b ( n ) = f b { x ( n ) } , b { 1 , , B } ,

Image (4.11)

as inputs for the linear B×1Image Multiple-Input/Single-Output (MISO) subsystem depicted in the right half of Fig. 4.3. This allows one to compute output samples

y ˆ ( n ) = b = 1 B h ˆ ( b ) , n 1 , x ˇ ( b ) , n ,

Image (4.12)

where hˆ(b),n1Image is the length-L IR estimate in branch b and has been obtained at time index n1Image and where xˇ(b),nImage is the time-reversed branch signal vector, structured analogously to Eq. (4.4). This allows one to compute the error signal

e ( n ) = y ( n ) y ˆ ( n )

Image (4.13)

before performing for each branch b a filter update according to

h ˆ ( b ) , n = h ˆ ( b ) , n 1 + μ b e ( n ) E x b ( n ) + δ x ˇ ( b ) , n ,

Image (4.14)

with the branch-specific adaptation step sizes μb(0,2)Image, a nonnegative regularization constant δ for numerical stability and the branch signal energies

E x b ( n ) = x ˇ ( b ) , n , x ˇ ( b ) , n .

Image (4.15)

The computational effort of such a time-domain NLMS-adaptive CGM grows linearly with the number of branches B. Note that the CGM adaptation includes the CM adaptation for B=1Image and the adaptation of linear systems for B=1Image and f1{x(n)}=x(n)Image.

4.3.1.2 Partitioned-Block Frequency-Domain NLMS

To efficiently adapt LIP nonlinear filters with a low input/output delay, Partitioned Block (PB) frequency-domain algorithms can be applied. For this purpose, a PB-FNLMS will be summarized here for CGMs.

Partitioned-Block Convolution

For partitioned-block convolution, the summation in Eq. (4.2) is split into partial sums and each partial sum (the convolution of an IR partition with an appropriately delayed segment of the input signal) is computed for M consecutive samples via fast DFT-domain convolution, like in [5862]. This leads to block processing of the input signal x(n)Image with a frameshift of M samples and a frame size of N=2MImage. In this notation, let the input signal vector xκImage at frame κ, the IR partition vector h(p)Image (containing the pth partition of h) and the output signal vector yκImage at frame κ be defined as

x κ = [ x ( κ M M ) , . . . , x ( κ M + M 1 ) ] T ,

Image (4.16)

h ( p ) = [ h ( p M ) , . . . , h ( p M + M 1 ) , 0 , . . . , 0 ] T and

Image (4.17)

y κ = [ 0 , . . . , 0 , y ( κ M ) , . . . , y ( κ M + M 1 ) ] T ,

Image (4.18)

respectively. Employing the length-N DFT matrix F, DFT-domain representations

x _ κ = F x κ     and 

Image (4.19)

h _ ( p ) = F h ( p )

Image (4.20)

can be obtained, and the output signal vector can be computed as

y κ = W 01 F H p = 0 P 1 x _ κ p h _ ( p ) y _ κ y κ ,

Image (4.21)

where y_κImage results from the accumulation over all P partitions' DFT-domain products and contains cyclic convolution artifacts in its time-domain representation yκImage. Thus, the time-domain output signal vector yκImage emerges from yκImage by premultiplication with the windowing matrix W01Image, defined in Eq. (4.5).

Application of Partitioned-Block Convolution for Adaptive Filtering

As in the previous paragraph's partitioned convolution scheme, an adaptive CGM estimate for the linear subsystem in branch b at frame κ can be split into partitions hˆ(b),κ(p)Image. Their DFT-domain representations hˆ_(b),κ(p)Image can be adapted efficiently on a frame-wise basis by an NLMS-type algorithm. To this end, branch signal vectors

x ( b ) , κ = [ [ 0 M , I M ] x ( b ) , κ 1 [ f b { x ( κ M ) } , , f b { x ( κ M + M 1 ) } ] T ]

Image (4.22)

are determined, which contain M samples of the previous frame (upper half) and M newly computed branch signal samples (lower half), and they are transformed into the DFT domain according to

x _ ( b ) , κ = F x ( b ) , κ .

Image (4.23)

Applying the fast partitioned convolution scheme of Eq. (4.21) in all B branches and summing up all branches in the DFT domain yields the output estimate with cyclic convolution artifacts

y _ ˆ κ = b = 1 B p = 0 P 1 x _ ( b ) , κ p h ˆ _ ( b ) , κ 1 ( p ) .

Image (4.24)

This result allows one to compute the DFT-domain error signal vector

e _ κ = F ( y κ W 01 F H y _ ˆ κ y ˆ κ ) ,

Image (4.25)

where yˆκImage represents the output signal vector estimate, obtained from y_ˆκImage after an Inverse DFT (IDFT) and windowing. For NLMS adaptation, the DFT-domain signal energy can be estimated using a recursive averaging

Image (4.26)

where Image is a smoothing factor (Image will be assumed as default) and will be used to compute normalized branch signals

x _ norm , ( b ) , κ = μ b ( x _ ( b ) , κ ) ( s x b , κ + δ ) ,

Image (4.27)

where μb(0,2)Image is the adaptation step size and δ is a small positive constant for numerical stability, which is added to each element of sxb,κImage. Based on x_norm,(b),κImage, the update of the filter partitions can be expressed as

h _ ˆ ( b ) , κ ( p ) = h ˆ _ ( b ) , κ 1 ( p ) + x _ norm , ( b ) , κ p e _ κ ,

Image (4.28)

which is also known as unconstrained update in the literature [58] and where cyclic convolution artifacts lead to nonzero filter taps in the second half of hˆ(b),κ(p)=FHh_ˆ(b),κ(p)Image. Limiting the temporal support of the partitions to M samples is possible by

h ˆ _ ( b ) , κ ( p ) = F W 10 F H ( h _ ˆ ( b ) , κ ( p ) ) h ˆ ( b ) , κ ( p )

Image (4.29)

and corresponds to the so-called constrained update in the literature, which is typically formulated by introducing the constraint on the update instead of the partitions (see [58]). Alternatively, a soft-partitioned update is also possible [62,63], which shapes the temporal support of hˆ(b),κ(p)Image by convolution in the DFT domain with a very short DFT-domain sequence in order to save the DFTs in Eq. (4.29).

Corresponding to the time-domain adaptation in Section 4.3.1.1, this section's PB-FNLMS CGM adaptation includes the CM adaptation for B=1Image and the adaptation of linear systems for B=1Image and f1{x(n)}=x(n)Image. Furthermore, a nonpartitioned Frequency-domain Normalized Least Mean Square (FNLMS) algorithm results from P=1Image. Taking the number of DFTs as measure of computational complexity, a CGM with B=5Image branches and P=4Image partitions is about 4.3 times as complex as a linear model with P=4Image partitions (B=1Image).

4.3.2 Filtered-X Adaptation for Bilinear Cascade Models

In this section, the FX algorithm will first be introduced and discussed on a conceptual level for bilinear CMs, before specifying it in detail and proposing an efficient realization for CGM preprocessors with single-tap branch filters (i.e., parametric preprocessors according to Eq. (4.8)).

The Generic Filtered-X Structure  FX algorithms are frequently employed in Active Noise Control (ANC) [64] for prefilter adaptation. Thereby, the FX algorithm exploits the fact that the ordering of linear time-invariant systems in a cascade can be reversed without altering the cascade's output: for the preprocessor-based CM in Fig. 4.4B, joint linearity and time invariance allow incorporating h into each of the branches, as depicted in Fig. 4.5. Therein, the prefiltered inputs to the former prefilters wbImage are often termed FX signals. The representation of the CM in Fig. 4.5 allows one to directly apply an NLMS-type adaptation algorithm for linear MISO systems from Section 4.3.1 to adapt wbImage, as their inputs are known and their combined output is directly matched to the target signal (in system identification: an unknown system's output). Mathematically, an FX LMS algorithm can also be derived directly like an LMS algorithm as stochastic gradient descent algorithm w.r.t. the partial derivative w.r.t. the prefilters. The exchange of the filtering orders results simply from an advantageous summation order in the gradient calculation. Note that adaptive filters for wbImage and h actually violate the time invariance required for exchanging the filter order. However, the time variance appears to be uncritical in practice, where FX algorithms have been employed for ANC for more than three decades [64,65].

Image
Figure 4.5 A classical FX structure for an HM with parametric preprocessor according to Eq. (4.8) as a result of exchanging the filtering order of h and wb in Fig. 4.4B (exploiting linearity and assuming time invariance), which forms the common core of a class of adaptive algorithms for nonlinear system identification [24].

Review of Known Algorithms for Filtered-X Adaptation of Bilinear Cascade Models  Many FX-like algorithms have been derived for the adaptation of bilinear LIP nonlinear systems but do not establish the link to the filtered-X algorithms and the exchange of filtering orders. The following description of these methods will highlight the fact that these algorithms can be viewed and implemented as FX algorithms.

In [24], the FX preprocessor adaptation was derived and applied to the adaptation of a polynomial preprocessor with fb{}=()bImage—a nonadaptive offline version of this mechanism was proposed in [23], which resembles the iterative version of [51] for AutoRegressive Moving Average (ARMA) models as linear submodels. More recent work on such cascade models considered a generalization of [24] by allowing longer IRs than a single tap [37,66] or considered piece-wise linear functions for fb{}Image in conjunction with a modified joint normalization of the linear and nonlinear filter coefficients [27]. In particular, [37] employs a power filter as preprocessor to the linear subsystem. The resulting algorithm corresponds to a time-domain FX algorithm. In [66], a particular CGM preprocessor is discussed, which corresponds to a Volterra filter in triangular representation (see, e.g., [19] for the triangular representation) with memory lengths of N1Image and N2Image for the kernels of orders 1 and 2, respectively, and no memory at all for higher-order Volterra kernels (becoming a simple memoryless preprocessor3 for these orders). Structures similar to this preprocessor also result from the so-called EVOLutionary Volterra Estimation (EVOLVE), which will be outlined later on in Section 4.6. While evaluating their algorithm for a memoryless preprocessor (N1=N2=0Image), [66] also evaluates an extension of [24] by adapting the linear subsystem with a frequency-domain NLMS with unconstrained update. The preprocessor is adapted, as in [24], by a Recursive Least Squares (RLS)-like algorithm based on the FX signals. Still, the two RLS descriptions differ as [66] employs the direct inversion of the involved correlation matrix, whereas [24] makes use of the alternative variant with a recursively computed inverse (see [57] for RLS-adaptive filters). Another extension of [24] is treated in [25] (in German), where the signal flow and benefit of a subband-AEC variant of [24] is discussed on a theoretical level.

Note that none of these approaches describes the preprocessor adaptation explicitly as an FX algorithm.4 As opposed to this, the remainder of this section introduces an FX PB-FNLMS algorithm for CMs with CGM preprocessors and tailors this algorithm to the special case of memoryless preprocessors according to Eq. (4.8).

Filtered-X Adaptation of Preprocessors of Partitioned-Block CMs  In the following, consider the structure of Fig. 4.4B as a single-tap CGM, for which DFT-domain branch signal vectors can be computed according to Eq. (4.23). Based on that, preprocessed DFT-domain input signal vectors can be determined by

x _ pp , κ = b = 1 B w ˆ b ( κ 1 ) x _ ( b ) , κ

Image (4.30)

and the DFT-domain FX signal vectors x_FX,(b),κImage (see Fig. 4.5 illustrating the FX signals) can be computed by partitioned convolution in a two-step procedure as

x _ FX , ( b ) , κ = p = 0 P 1 x _ ( b ) , κ p h ˆ _ κ 1 ( p )

Image (4.31)

x _ FX , ( b ) , κ = F [ [ 0 M , I M ] x FX , ( b ) , κ 1 [ 0 M , I M ] F H x _ FX , ( b ) , κ ] x FX , ( b ) , κ .

Image (4.32)

With the introduced signals, the adaptive filter's output can alternatively be determined by one of the two equations

y ˆ κ = W 01 F H p = 0 P 1 x _ pp , κ p h ˆ _ p ( κ )

Image (4.33)

W 01 F H b = 1 B w ˆ b ( κ 1 ) x _ FX , ( b ) , κ y _ κ ,

Image (4.34)

where Eq. (4.33) corresponds to block processing with the signal flow in Fig. 4.4B and Eq. (4.34) corresponds to the signal flow in Fig. 4.5. For time-invariant systems, Eqs. (4.33) and (4.34) are identical. This allows one to compute the error signal vector eκ=yκyˆκImage and its DFT

e _ κ = F e κ ,

Image (4.35)

both of which will be used for NLMS-type adaptation of the parameters. Analogously to Section 4.3.1.2, a PB-FNLMS adaptation of hˆ_κ(p)Image can be expressed as

h _ ˆ κ ( p ) = h ˆ _ κ 1 ( p ) + x _ norm , pp , κ p e _ κ ,

Image (4.36)

h ˆ _ κ ( p ) = F W 10 F H ( h _ ˆ κ ( p ) ) h ˆ κ ( p ) ,

Image (4.37)

where x_norm,pp,κImage is the normalized preprocessed input signal vector computed analogously to Eqs. (4.26) and (4.27) (containing the adaptation step size) and where Eq. (4.36) and Eq. (4.37) correspond to the unconstrained and the constrained update, respectively.

Similar to [24], the weights wˆb(κ)Image are estimated by an NLMS algorithm employing the FX signals as

w ˆ b ( κ ) = w ˆ b ( κ 1 ) + μ FX E b + δ x FX , ( b ) , κ , e κ

Image (4.38)

with the adaptation step size μFXImage and

E b = x FX , ( b ) , κ , x FX , ( b ) , κ .

Image (4.39)

For the more general case where a partitioned-block CGM with PppImage partitions hˆ_(b),κ(p)Image replaces the preprocessor with weights wˆb(κ)Image, all linear combinations with the weights, like in Eq. (4.30), have to be replaced by actual partitioned convolutions and PB-FNLMS updates can be computed for the preprocessor according to

h _ ˆ ( b ) , κ ( p ) = h ˆ _ ( b ) , κ 1 ( p ) + x _ FX , norm , ( b ) , κ p e _ κ

Image (4.40)

and

h ˆ _ ( b ) , κ ( p ) = F W 10 F H ( h _ ˆ ( b ) , κ ( p ) ) h ˆ ( b ) , κ ( p )

Image (4.41)

with the normalized DFT-domain FX branch signal vectors x_FX,norm,(b),κpImage determined from the FX branch signals xFX,(b),κImage analogously to Eqs. (4.26) and (4.27). This corresponds to a generalization of [37] from time-domain power filter preprocessor adaptation to PB-FNLMS adaptation of general CGM preprocessors.

Thereby, the adaptation of bilinear CMs has been decomposed into two ordinary PB-FNLMS system identification tasks, plus the additional prefiltering operations generating the FX branch signals. All these components can be implemented efficiently using partitioned-block convolution techniques.

An Adaptation Tailored to Single-Tap CGM Preprocessors

In the following, the implications of a parametric preprocessor according to Eq. (4.30) (a single-tap CGM preprocessor, cf. Fig. 4.4B) will be considered in detail. Interestingly, Eq. (4.34) does not introduce cyclic convolution artifacts into y_κImage for single-tap filters wˆb(κ)Image. Consequently, the cyclic convolution artifacts present in xFX,(b),κImage do not spread by filtering with the weights, allowing to compute the output estimate

y ˆ κ = W 01 F H b = 1 B w ˆ b ( κ 1 ) x _ FX , ( b ) , κ

Image (4.42)

without Eq. (4.32). Furthermore, the zeros in the vector eκImage and the unitarity of the DFT imply

x FX , ( b ) , κ , e κ = x FX , ( b ) , κ , e κ

Image (4.43)

= x _ FX , ( b ) , κ , e _ κ .

Image (4.44)

Further identifying

E b E b = x _ FX , ( b ) , κ , x _ FX , ( b ) , κ

Image (4.45)

as a reasonable approximation of the branch signal energy within a frame allows one to compute the weights according to

w ˆ b ( κ ) = w ˆ b ( κ 1 ) + μ FX E b + δ x _ FX , ( b ) , κ , e _ κ

Image (4.46)

without Eq. (4.32) as well. Thus, Eq. (4.32) does not need to be evaluated at all, which allows saving 2B DFTs of length N.

The remaining DFTs and IDFTs are B+2Image transforms for computing DFT-domain branch signals prior to Eq. (4.31), the output estimate in Eq. (4.34), and the DFT-domain error signal in Eq. (4.35), as well as 2P DFTs for a constrained update of the partitioned linear subsystem in Eq. (4.37). The computational effort for non-DFT operations has a component growing proportional to BPImage due to Eq. (4.31). However, as the effort for a length-N DFT is typically much higher than the effort for a vector multiplication in Eq. (4.31), the number of DFTs may serve as a first indicator of the computational complexity. Thus, the relative computational complexity RCGMFX-CMImage of an FX CM and a full-length CGM can be approximated by

R ˜ CGM FX-CM = B + 2 + 2 P B + 2 + 2 B P + ε ,

Image (4.47)

which is the ratio of the numbers of involved DFTs plus an overhead ε representing the number of operations which are not DFTs and grow with BPImage in both the FX-CM and the CGM (filtering and filter coefficient updates). To give an example, P=4Image and B=5Image leads to R˜CGMFX-CM0.32+εImage and suggests that the CM can be identified with about one-third of the effort of a CGM of the same length.

4.3.3 Summary of Algorithms

The main algorithms presented in Section 4.3 are summarized in tabular form to support an implementation in practice. The most basic algorithm is the time-domain NLMS algorithm summarized in Table 4.1. The operations listed therein have to be evaluated at each sample index n. For processing an entire block of M subsequent samples, as in block- and frequency-domain processing, all equations in Table 4.1 have to be evaluated sample by sample for each of the block's samples—M times in total. For PB-FNLMS adaptation, the operations for processing a block of M samples are summarized in Table 4.2. The operations listed therein have to be executed for each signal frame κ (once every M samples). In the same way, the operations required for FX adaptation of a bilinear CM with a single-tap CGM as nonlinear preprocessor are listed in Table 4.3.

Table 4.1

Summary of operations of a time-domain NLMS algorithm for CGMs as executed for each sampling instant n

computed quantity computation formula equation no.
branch signals x b ( n ) = f b { x ( n ) } Image b Eq. (4.11)
time-reversed branch signal vectors x ˇ ( b ) , n = [ x b ( n ) , , x b ( n L + 1 ) ] T Image b
output signal y ˆ ( n ) = b = 1 B h ˆ ( b ) , n 1 , x ˇ ( b ) , n Image Eq. (4.12)
error signal e ( n ) = y ( n ) y ˆ ( n ) Image Eq. (4.13)
frame energy E x b ( n ) = x ˇ ( b ) , n , x ˇ ( b ) , n Image b Eq. (4.15)
filter coefficients h ˆ ( b ) , n = h ˆ ( b ) , n 1 + μ b e ( n ) E x b ( n ) + δ x ˇ ( b ) , n Image b Eq. (4.14)

Image

Table 4.2

Summary of operations of a PB-FNLMS algorithm for CGMs as executed for each frame κ

computed quantity computation formula equation no.
branch signals x ( b ) , κ = [ [ 0 M , I M ] x ( b ) , κ 1 [ f b { x ( κ M ) } , , f b { x ( κ M + M 1 ) } ] T ] Image b Eq. (4.22)
∼ in DFT domain x _ ( b ) , κ = F x ( b ) , κ Image b Eq. (4.23)
DFT-domain output estimate with cyclic convolution artifacts y _ ˆ κ = b = 1 B p = 0 P 1 x _ ( b ) , κ p h ˆ _ ( b ) , κ 1 ( p ) Image Eq. (4.24)
DFT-domain error signal e _ κ = F ( y κ W 01 F H y _ ˆ κ ) Image Eq. (4.25)
branch PSDs Image b Eq. (4.26)
normalized branch signals x _ norm , ( b ) , κ = μ b ( x _ ( b ) , κ ) ( s x b , κ + δ ) Image b Eq. (4.27)
updated filters h _ ˆ ( b ) , κ ( p ) = h ˆ _ ( b ) , κ 1 ( p ) + x _ norm , ( b ) , κ p e _ κ Image b,p Eq. (4.28)
constrained update h ˆ _ ( b ) , κ ( p ) = F W 10 F H ( h _ ˆ ( b ) , κ ( p ) ) Image b,p Eq. (4.29)

Image

Table 4.3

Summary of operations of an FX algorithm tailored to the adaptation of bilinear CMs with single-tap CGM preprocessors as executed for each frame κ

computed quantity computation formula equation no.
branch signals x ( b ) , κ = [ [ 0 M , I M ] x ( b ) , κ 1 [ f b { x ( κ M ) } , , f b { x ( κ M + M 1 ) } ] T ] Image b Eq. (4.22)
∼ to DFT domain x _ ( b ) , κ = F x ( b ) , κ Image b Eq. (4.23)
DFT-domain preprocessed signal x _ pp , κ = b = 1 B w ˆ b ( κ 1 ) x _ ( b ) , κ Image Eq. (4.30)
DFT-domain FX branch signals with cyclic convolution artifacts x _ FX , ( b ) , κ = p = 0 P 1 x _ ( b ) , κ p h ˆ _ κ 1 ( p ) Image b,p Eq. (4.31)
output signal y ˆ κ = W 01 F H b = 1 B w ˆ b ( κ 1 ) x _ FX , ( b ) , κ Image Eq. (4.33)
error signal e κ = y κ y ˆ κ Image
∼ to DFT domain e _ κ = F e κ Image Eq. (4.35)
preprocessed signal PSD Image
normalized input x _ norm , pp , κ = ( x _ pp , κ ) ( μ b ( s x pp , κ + δ ) ) Image
updated filters h_ˆκ(p)=hˆ_κ1(p)+x_norm,pp,κpe_κImage, p Eq. (4.40)
constrained update hˆ_κ(p)=FW10FH(h_ˆκ(p))hˆκ(p)Image, p Eq. (4.41)
FX energies E b E b = x _ FX , ( b ) , κ , x _ FX , ( b ) , κ Image b > 1 Eq. (4.45)
weight update w ˆ b ( κ ) = w ˆ b ( κ 1 ) + μ FX E b + δ x _ FX , ( b ) , κ , e _ κ Image b > 1 Eq. (4.46)

Image

These algorithms will be employed as building blocks of the computationally very efficient SA filters in the following section, Section 4.4. Therein, the aforementioned basic algorithms will only be referenced without repeating all equations.

4.4 Significance-Aware Filtering

In [28], the concept of SA filtering has been introduced, in order to estimate the parameters of nonlinear CMs by decomposing the adaptation process in a divide-and-conquer manner into synergetic adaptive subsystems [28,3032] which are LIP. Thereby, the estimation of the nonlinearity is separated from the estimation of a possibly long linear subsystem, as it would be characteristic for acoustic echo paths. As a consequence, the nonlinearity can be estimated computationally efficiently by nonlinear models with a very low number of parameters.

In the following, two different SA decompositions will be introduced in Section 4.4.1 and employed for the Serial SA CMs (SSA-CMs) in Section 4.4.2 (a generalization of the Equalization-based SA (ESA)-HM from [32]), the Parallel SA CGMs (PSA-CGMs) in Section 4.4.3 (a generalization of the SA-HGM from [32]) and a novel Parallel SA Filtered-X (PSAFX) algorithm in Section 4.4.4.

4.4.1 Significance-Aware Decompositions of LIP Nonlinear Systems

In [32], two different SA decompositions have been employed, which will be denoted Serial SA (SSA) decomposition (employed for ESA filtering in [32]) and Parallel SA (PSA) decomposition (employed for so-called “classical” SA filtering in [32]). Both decompositions allow for the estimation of the nonlinearity with a low-complexity nonlinear adaptive filter, as will be explained in the following.

Serial Significance-Aware Decomposition

The SSA decomposition is depicted schematically in Fig. 4.6, where the unknown system is cascaded with an equalizer filter hˆeqImage. The overall response of this serial connection will ideally be a (delayed) version of the nonlinear preprocessor f{}Image. Thus, an estimate for the physically inaccessible intermediate signal xNL(n)Image is obtained, which allows one to estimate the nonlinearity directly from this decomposition efficiently by a nonlinear model with a short temporal support. An adaptive system using this decomposition will be specified in Section 4.4.2.

Image
Figure 4.6 Serial decomposition of the unknown system exploited for SSA filtering. Cascading the unknown system with a linear equalizer filter allows one to approximate the cascade as the nonlinear component of the unknown system (disregarding the processing delay due to causal filtering).

Parallel Significance-Aware Decomposition

The PSA decomposition is illustrated in Fig. 4.7. The LIP property of a nonlinear CM is employed to decompose its IR vector into a dominant region hdImage (light gray), describing, e.g., direct-path propagation, and a complementary region hcImage (dark gray). Augmenting the unknown system with a parallel system modeling the complementary-region output signal component yc(n)Image allows one to compute the dominant-region output component yd(n)=y(n)yc(n)Image. As f{}Image, xNL(n)Image, hcImage, and thereby yc(n)Image are not known during system identification, they have to be replaced by estimates fpp{}Image, xpp(n)Image, hˆcImage and yˆc(n)Image in practice, as depicted in the upper part of Fig. 4.7. Still, the output of the parallel arrangement yields an estimate ec¯(n)=y(n)yˆc(n)yd(n)Image of the dominant-region component of the unknown system (bottom of Fig. 4.7). As the dominant-region component of the system describes the transmission of a significant amount of energy, a nonlinear dominant-region model can be estimated at a high Signal to Noise Ratio (SNR) and will suffer much less from gradient noise than a reverberation tail model. Adaptive systems using this decomposition will be specified in Sections 4.4.3 and 4.4.4.

Image
Figure 4.7 Parallel decomposition of the unknown system exploited for PSA filtering. The temporal support modeled by IR vectors is illustrated below the respective vectors' symbols as follows. The entire vector h is represented by a flat white box with a gray frame. Nonzero entries are marked by light gray boxes if they correspond to the dominant region and by dark gray boxes if they describe the complementary region.

Estimating a Preprocessor System From a CGM

Another key component for the SA models described in Sections 4.4.2 and 4.4.3 below is the possibility to extract a CM's preprocessor from an identified CGM-structured estimate of the CM system. To this end, assume that the CM's nonlinearity can be expressed as weighted sum of the CGM's branch nonlinearities fb{}Image, resulting in an expression identical to Eq. (4.8). The CM's linear subsystem will be denoted h. Then, the CM can be expressed as CGM with linear subsystems h(b)=wbhImage. In practice, an identified CGM can only provide noisy estimates

h ˆ ( b ) = w b h + ϵ b ,

Image (4.48)

where ϵbImage is the coefficient error vector. As shown in [28], a least squares estimate for wbImage can be obtained by

w ˆ b ( LS ) = h ( b ) , h ( b ref ) h ( b ref ) , h ( b ref )

Image (4.49)

w.r.t. a reference branch h(bref)Image, for which wˆbref(LS)=1Image. Note that wˆbref(LS)=1Image is not a model restriction—it simply removes the scaling ambiguity inherent to the cascade model by shifting the actual wbrefImage as gain factor into the estimate hˆImage to be identified after the preprocessor. Without loss of generality, bref=1Image and f1{}=()Image will be assumed from now on. A preprocessor with wb=0b>1Image will be referred to as linearly configured preprocessor.

4.4.2 Serial Significance-Aware Cascade Models

SSA-CMs employ the SSA decomposition from Fig. 4.6 and generalize the Serial SA HMs (SSA-HMs) from [32] by allowing branch nonlinearities with memory. The structure of an SSA-CM is depicted in Fig. 4.8. Initially, the preprocessor in Block A in Fig. 4.8, providing the signal xpp(n)Image, is configured to be linear. After this initialization, all subsystems in Fig. 4.8 are adapted in parallel. In Block C in Fig. 4.8, an adaptive linear equalizer implements the SSA decomposition. The equalizer filters the unknown system's output y(n)Image to produce an estimate xˆNL(nNeq)Image of the delayed nonlinearly distorted input signal. For adaptation, xˆNL(nNeq)Image is matched to xpp(nNeq)Image.5 Ideally, x(nNeq)Image and xˆNL(nNeq)Image are time-aligned as well and related by f{}Image (cf. Fig. 4.6). Thus, the nonlinear relationship between x(nNeq)Image and xˆNL(nNeq)Image can be estimated by a very short adaptive CGM of L=LSAImage taps in Block B in Fig. 4.8. Although LSA=1Image would be sufficient in the case of a perfect equalization of h, a very small LSAImage of about LSA=3Image is reasonable in practice due to the adaptive, imperfect equalizer. Between Blocks B and A in Fig. 4.8, the identified CGM and Eq. (4.49) are employed to estimate the coefficients of a preprocessor which combines the CGM's branch signals according to Eq. (4.8). Analogously to [32], the preprocessor coefficients wˆbImage are estimated on a frame-wise basis as

w ˆ b ( κ ) = γ w w ˆ b ( κ 1 ) + ( 1 γ w ) w ˜ b ( κ ) ,

Image (4.50)

where γw,0γw<1Image, is a recursive-smoothing constant (γw=0.95Image will be used by default for the simulations in Section 4.5) and w˜b(κ)Image is computed according to Eq. (4.49) from the instantaneous estimates of the branch signal vectors of the CGM of Block B. The preprocessor with weights wˆb(κ)Image is used in a CM with an adaptive linear subsystem to model the entire unknown system (Block A in Fig. 4.8).

Image
Figure 4.8 Proposed novel SSA filtering: an estimate xˆNL(n)Image of the nonlinearly distorted input signal is obtained in Block C by equalizing the system output signal y(n)Image with a linear equalizer, which allows in Block B to assess the nonlinear distortions by a very short CGM, which in turn is employed for estimating the parameters of the nonlinear preprocessor fpp{}Image of Block A.

The SSA-CM structure according to Fig. 4.8 is efficient, because it requires only two adaptive linear filters with long temporal support (in Blocks A and C), both of which will be assumed to have length L, and a CGM with a very low number of taps LSAImage. A number of LSA=3Image taps will be assumed by default in the following. The long adaptive filters can be realized efficiently using a PB-FNLMS according to Section 4.3.1.2 in a block processing scheme and the CGM can be implemented due to its short length at very low computational effort by a time-domain NLMS according to Section 4.3.1.1. Hence, the computational effort for an SSA-CM is roughly twice as high as that for a linear filter. The relative complexity compared with a CGM, RCGMSSA-CMImage, is about

R ˜ CGM SSA-CM = 2 P B P = 2 B ,

Image

where the contribution of the length-LSAImage CGM has been disregarded. For B=5Image branches and P=4Image partitions, this leads to R˜CGMSSA-CM=0.40Image.

4.4.3 Parallel Significance-Aware Cascade Group Models

PSA-CGMs generalize the SA-HGMs from [28,31] by allowing branch nonlinearities with memory and are structurally identical otherwise. The structure of such a PSA-CGM is depicted in Fig. 4.9. For an explanation of the general concept, assume that the temporal support of the unknown system's dominant region (cf. Fig. 4.7) is known. In this case, all adaptive systems in Fig. 4.9 can be adapted simultaneously.

Image
Figure 4.9 Basic structure of a PSA-CGM: a B-branch CGM models only the dominant-part propagation (Block C), whereas the long IR is covered by a simple preprocessor system (Block A).

Block A of Fig. 4.9 contains a CM with adaptive linear subsystem (minimizing the error signal eCM(n)Image). The nonlinear parametric preprocessor fpp{}Image is initially configured as linear function and is refined separately from the IR. The CM's IR estimate is partitioned into a dominant region hˆdImage (upper branch in Block A) and a complementary region hˆcImage (lower branch in Block A). This allows one to apply the PSA decomposition (cf. Fig. 4.7) to the unknown system in Block B in Fig. 4.9: subtracting the complementary-region estimate yˆc(n)Image from the unknown system's output y(n)Image results in the dominant-region dominated error signal ec¯(n)Image. This enables in Block C of Fig. 4.9 the identification of the unknown system's dominant region alone by a CGM with short temporal support (as short as hˆdImage). The CGM's IRs hˆ(b)Image allow one to extract the preprocessor coefficients for the CM in Block A of Fig. 4.9 analogously to Section 4.4.2. It is worth noting that a PSA-CGM offers two different error signals: eCM(n)Image, resulting from a CM output signal estimate yˆCM(n)Image, and eSA(n)Image, resulting from yˆSA(n)=yˆc(n)+yˆCGM(n)Image (not shown in Fig. 4.9). The estimate yˆSA(n)Image is obtained with a CGM as dominant-path model and represents a more general model than the CM in Block A of Fig. 4.9. If the unknown system has a CM structure, yˆCM(n)Image may be slightly more accurate than yˆSA(n)Image, because the CGM has more degrees of freedom than necessary and is more affected by gradient noise. If the unknown system is more complex than a CM, the additional degrees of freedom of the CGM may render yˆSA(n)Image more accurate. For this reason, the PSA-CGM will appear in the evaluation in Section 4.5 twice.

For a highly efficient low-delay realization of the PSA-CGM, a PB-FNLMS implementation has been proposed in [31]. This method identifies partitions hˆκ(p)Image of the linear subsystem of the CM in Block A of Fig. 4.9 by a PB-FNLMS (Section 4.3.1.2, B=1Image branches and P>1Image partitions). Employing this partitioning, the dominant region is modeled by a single partition hˆκ(pd)Image with index pdImage and the complementary region is composed of all other partitions, where ppdImage. As a consequence, the nonlinear dominant-region model of Block C in Fig. 4.9 is a single-partition CGM and adapted according to Section 4.3.1.2 (B>1Image branches and P=1Image partition).

If the temporal support of the dominant region is unknown, only hˆκ(p)Image are adapted in an initial convergence phase. Afterwards, the fixed-length partitions hˆκ(p)Image allow one to detect a dominant partition hˆκ(pd)Image as the partition with maximum energy. Compared to a full-length CGM with P partitions in each branch, the relative computational complexity RCGMPSA-CGMImage of a PSA-CGM is approximately

R ˜ CGM PSA-CGM = B + P B P .

Image

Obviously, the PSA-CGM is very efficient for large P and B. Without block partitioning (P=1), however, the described PSA-CGM does not provide any advantage. To give an example, B=5Image and P=4Image results in R˜CGMPSA-CGM=0.45Image.

4.4.4 Parallel Significance-Aware Filtered-X Adaptation

In this section, a novel SA filtering concept will be introduced. This concept exploits the PSA decomposition for the direct FX adaptation of bilinear CMs from Section 4.2.3 (see Fig. 4.7) and will be denoted Parallel SA Filtered-X CM (PSAFX-CM). The block diagram of the PSAFX-CM matches the block diagram of the PSA-CGM in Fig. 4.9, except for Block C. Like for the PSA-CGM, a CM with an adaptive linear subsystem (Block A in Fig. 4.9) enables a PSA decomposition (Block B in Fig. 4.9); in an initialization phase, the CM preprocessor is configured as linear function and only the IR is adapted. Afterwards, the CM's IR is split into a dominant and a complementary region. The complementary-region submodel is used (as in Block B in Fig. 4.9) to compute an error signal ec¯(n)=y(n)yˆc(n)yd(n)Image. This error signal mainly contains components yd(n)Image, which are caused by the unknown system's dominant region (see also Fig. 4.7).

As an alternative to the PSA-CGM, the novel PSAFX-CM implements the adaptive nonlinear model for the dominant region (Block C) according to Fig. 4.10. Therein, the PSAFX-CM employs a dominant-region CM with hˆdImage from Block A in Fig. 4.9 and FX-adapted preprocessor weights wˆbImage. As it is characteristic for SA filtering, the preprocessor determined for the dominant region is then applied in the CM with the full-length adaptive linear subsystem (Block A in Fig. 4.9). Note that the complementary-region CM from Block A in Fig. 4.9 and the dominant-region CM from Fig. 4.10 seamlessly combine to the overall CM of Block A in Fig. 4.9 again. This implies eFX(n)=eCM(n)Image. Thus, the only difference from an ordinary FX implementation is that the FX branch signals are generated with the most significant fragment of hˆImage.

Image
Figure 4.10 Substitution for Block C in Fig. 4.9 to form the novel PSAFX-CM: The preprocessor coefficients wˆbImage of a dominant-region CM are adapted using an FX algorithm with hˆdImage from Block A in Fig. 4.9 as linear subsystem of the CM and with ec¯(n)yd(n)Image from the PSA decomposition in Block B in Fig. 4.9 as target signal.

Pursuing a PB-FNLMS implementation as in Section 4.3.2, the adaptation procedures are identical up to Eq. (4.31), and Eq. (4.31) simplifies to the single partition

x _ FX , ( b ) , κ = x _ ( b ) , κ p h ˆ _ κ 1 ( p d ) = : x _ SAFX , ( b ) , κ

Image (4.51)

and is therefore independent of the length of the actually modeled unknown system. Unlike for the FX algorithm of Section 4.3.2, the filtering effort for generating the FX signals is not proportional to BPImage anymore, but only to B. As a consequence, the relative complexity RCGMPSAFX-CMImage of a PSAFX-CM w.r.t. a full-length CGM can be estimated by the number of DFTs and yields

R ˜ CGM PSAFX-CM = B + 2 + 2 P B + 2 + 2 B P < R ˜ CGM FX-CM .

Image (4.52)

The overhead ε from Eq. (4.47), caused by non-DFT operations with a complexity proportional to BPImage, does not appear in Eq. (4.52) as the actual filtering effort is reduced to a complexity proportional to B only. To give an example, P=4Image and B=5Image leads to R˜CGMPSAFX-CM0.32Image and thereby indicates a computationally very inexpensive nonlinear model, which is even less complex than the other SA filters.

As for the other SA models in Sections 4.4.2 and 4.4.3, the nonlinear system identification has been split into synergetic subsystems, where the first models the possibly long memory of the linear subsystem (Block A in Fig. 4.9) and the second estimates the nonlinearities without modeling the possibly long memory of the system (Fig. 4.10).

4.5 Experiments and Evaluation

In this section, the adaptive LIP nonlinear filters from Sections 4.3 and 4.4 will be evaluated exemplarily for the challenging application of AEC (cf. Section 4.1). To this end, the evaluation method will be introduced in Section 4.5.1 and the experiments will be presented and discussed in Section 4.5.2.

Note that, in principle, the presented methods can also be employed for many applications aside from AEC. In particular, the identification of the path between the discrete-time loudspeaker signal and a recording of the radiated sound can also be the first step towards a loudspeaker linearization [79] or nonlinear active noise control [68,53]. Furthermore, the presented algorithms can enable the identification of nonlinear analog signal processors (e.g., guitar amplifier and effect processor devices [69,70]) to emulate the identified processors in digital hardware afterwards. Moreover, the joints of robots can be modeled as HMs as well [71] and may be identified with the methods of Section 4.3. Lifting the latter example to a broad industrial scope, the adaptive algorithms presented in this chapter may be employed for digital twin modeling (also termed cybertwin) of nonlinear mechanical, electrical, acoustical or chemical systems in digitalized production plants for Industry 4.0 [72]—as long as the underlying nonlinear systems can be described well by CMs. However, the longer the linear subsystem of an HM or CM is, the more computational benefit can be expected by SA modeling. In case of short linear subsystems, block partitioning and frequency-domain processing may be unnecessary and time-domain adaptation of all involved adaptive linear filters should be pursued (see [28] for a time-domain SA filter).

4.5.1 Evaluation Metrics

For system identification, normalized misalignment [24] and projection misalignment measures [73] evaluate how well particular components of a parametrized system are identified. However, these measures require the exact knowledge of the true system parameters—knowledge which is typically not available for actual physical systems to be identified. In this case, signal-based measures have to be adopted. In particular, the modeling accuracy can be measured by the ratio of the variance of the unknown system's output y(n)Image and of the residual error e(n)=y(n)yˆ(n)Image.

For the special system identification case of AEC, where the primary objective is the removal of the loudspeaker signal components (echoes) from the microphone signal, this ratio of variances is widely known as Echo-Return Loss Enhancement (ERLE) and is usually considered in the logarithmic domain according to

ERLE = 10 log 10 ( E { | y ( n ) | 2 } E { | e ( n ) | 2 } ) dB ,

Image (4.53)

assuming noise-free “single talk” situations (no speaker or noise on the near end, only an echo signal). A higher ERLE measure corresponds to a better performance of the AEC system [74]. When estimating the ERLE in short time frames employing the instantaneous energies of y(n)Image and e(n)Image and continuously adapting the model, even severe overadaptation to the particular signal may remain unnoticed and be misinterpreted as accurate system identification because of the high ERLE values. To minimize this undesired effect, filtering a long interval of speech with frozen filter coefficients after convergence is possible. The resulting ERLE can be used as system identification performance measure. An ERLE computed in this way also evaluates the primary use case of an AEC system, as it captures the amount of echo reduction in the presence of near-end speech (double talk), when the adaptive filters cannot be adapted and the recording cannot be muted (no half-duplex but full-duplex communication) [74]. The ERLE measure for frozen filter coefficients will be employed in the following to evaluate and compare the adaptive filtering algorithms described in Sections 4.3 and 4.4.

4.5.2 Experiments

In this section, the ERLE performance of the adaptive algorithms from Sections 4.3 and 4.4 will be compared in AEC experiments. To this end, the nonlinear echo paths of different devices will be considered. Device 1 is a smartphone in hands-free mode, Device 2 is a low-quality electrodynamic loudspeaker and Device 3 is a cellphone manufactured in the year 2000. About 80sImage of speech (both male and female speech) have been played back at relatively high volumes of about 80dBAImage at 1mImage distance and recorded with the playback device itself for Device 1 and with a high-quality measurement microphone at 1mImage distance for Devices 2 and 3. All AEC experiments were conducted at a sampling rate of fS=16kHzImage and in three acoustic environments, denoted Scenarios 1, 2 and 3. The actual recordings were performed under low-reverberation conditions, which corresponds to Scenario 1. The data for Scenarios 2 and 3 are synthesized from the low-reverberation recordings by convolving the recordings with measured acoustic IRs from real lab environments with reverberation times of T60250msImage and T60400msImage, respectively. The adaptive filters employed for the identification of the acoustic echo path will be realized, as specified earlier in this chapter, as partitioned-block adaptive filters in the frequency domain. For all scenarios and devices, a linear filter (Section 4.3.1.2 with B=1Image branches and f1{}=()Image), a CM with FX adaptation tailored to memoryless preprocessors (Section 4.3.2), a CM with PSAFX adaptation (Section 4.4.4), a PSA-CGM (Section 4.4.3), an SSA-CM (Section 4.4.2) and a CGM with full temporal support (Section 4.3.1.2) will be compared. All adaptive filters are identically parametrized by step sizes of μb=μFX=0.1Image, block processing is done at a frame shift of M=512Image samples and a frame size of N=2MImage and the block-partitioned filters consist of P=4Image partitions of length M. Furthermore, the equalizer delay for the SSA-CM is chosen as Neq=2MImage samples. All nonlinear models have B=5Image branches and preprocessor coefficients, respectively, and employ odd-order Legendre polynomials of orders 1 to 9 as branch nonlinearities fb{}Image. Thereby, all CGMs are HGMs and all CMs are HMs. The computational complexity of the individual models relative to the complexity of a CGM is listed in Table 4.4. Therein, the first line (“approximate”) corresponds to the relative complexity estimates introduced with the algorithms in Sections 4.3 and 4.4. The second line (“FLOPS”) contains the relative numbers of FLoating Point Operations (FLOPS) of the particular method in comparison to a CGM. To this end, the FLOPS have been determined analogously to [32]. Operations counted as FLOPS are real-valued additions, multiplications and divisions. The complex-valued DFT-domain operations are considered by mapping complex-valued additions, multiplications and divisions to multiple real-valued operations. In particular, a complex-valued addition is counted as 2 real-valued FLOPS, a complex-valued multiplication is counted as 6 real-valued FLOPS and the division of a complex number by a real-valued number is counted as 2 real-valued FLOPS. Accumulating the FLOPS of all equations necessary for processing a signal frame of M samples by a particular filter structure corresponds to the algorithm's number of FLOPS. Normalizing these FLOPS to the FLOPS of a CGM leads to the relative FLOPS in line two of Table 4.4.

Table 4.4

Computational complexity of linear and nonlinear models relative to a CGM with B = 5 branches and P = 4 partitions. The approximate values are a summary of the numbers in the individual algorithms' sections, whereas the number of FLOPS are computed analogously to [32] for a frameshift of M = 512

linear PSAFX-CM FX-CM SSA-CM PSA-CGM CGM
approximate 0.23 0.32 0.32 + ε 0.40 0.45 1
FLOPs 0.23 0.32 0.37 0.50 0.58 1

Image

Obviously, the qualitative rating of the different algorithms in Table 4.4 is similar by both complexity estimation methods: the further right an algorithm is in Table 4.4, the more computationally demanding it is.

The ERLE achievable with the different adaptive filter structures with frozen parameters (like during double talk) is listed in Figs. 4.11A to 4.11C for Devices 1 to 3, respectively. Obviously, linear AEC (A in Fig. 4.11) suffers a lot from the severe nonlinearities, leading to ERLE below 10dBImage under all reverberation conditions and to particularly bad performance under the low-reverberation conditions of Scenario 1. All nonlinear models outperform the linear model for all devices and scenarios. The CGM (G in Fig. 4.11), being the most complex model, shows consistently high performance. However, in particular the PSA-CGM (E and F in Fig. 4.11) approaches or even outperforms the more complex CGM. Especially the PSA-CGM's CM output (E in Fig. 4.11) performs very well. This confirms that the actual acoustic echo paths can be approximated well by CMs. However, the FX adaptation of a CM (C in Fig. 4.11) closely approaches the performance of the CGM only for Device 1. Still, introducing further efficiency into the FX adaptation by a PSA decomposition (B in Fig. 4.11) does not lead to any noticeable disadvantage. The SSA-CM performance (D in Fig. 4.11) resembles the performance of other nonlinear approaches for Device 1, but performs worse for the other devices (while still outperforming the linear filter).

Image
Figure 4.11 ERLE performance: the efficient SA-filtering approaches allow for nonlinear AEC and may even reach the performance of the filters without the SA extension (SAFX-CM/FX-CM and SA-CGM/CGM). (A) Smartphone. (B) Electrodynamic Loudspeaker. (C) Cellphone.

To sum up, all SA filters successfully alleviate the computational burden for nonlinear AEC. The performance of a general CGM can be reached in the experiments by an efficient PSA-CGM. An even less complex (but also less effective) direct CM adaptation by an FX algorithm can benefit in terms of computational complexity from PSA filtering without losing ERLE performance.

4.6 Outlook on Model Structure Estimation

All methods described in this chapter identify the parameters of nonlinear models adaptively. Yet, the algorithms do not choose an optimum model structure, such as a particular base function set, the order of a polynomial nonlinearity, or the memory length of linear subsystems. However, knowledge about the model structure is highly desirable for two reasons. First, minimizing the number of parameters leads to computational efficiency and second, unnecessarily estimated and redundant parameters introduce gradient noise (estimation errors) and consequently reduce the accuracy of the model.

Both disadvantages can be tackled by employing convex filter combinations [75], as described in Chapter 11. To briefly illustrate this concept for the scope of this chapter, consider two competing models A and B with the respective outputs yˆ(A)(n)Image and yˆ(B)(n)Image, modeling the same system with output y(n)Image. Although modeling the same system, the models A and B are assumed to have a different structure (e.g., memory length of linear subsystems) and consequently different numbers and values of parameters (e.g., filter coefficients of linear subsystems). Convex filter combinations superimpose the outputs y(A)(n)Image and y(B)(n)Image according to

y ˆ ( n ) = ( 1 η ) y ˆ ( A ) ( n ) + η y ˆ ( B ) ( n ) ,

Image

where the mixing weight η,0η1Image, is determined adaptively. In practice, this can perform like an adaptive switch indicating the more suitable of the two competing models and thereby the more suitable model structure. Convex combination-based methods for structure estimation can be applied to all kinds of CGMs (e.g., Volterra, Legendre, Chebyshev and Fourier nonlinear filters). As the structure estimation methods are described for the example of Volterra filters in the original publications, the methods will be explained here first for Volterra filters. To this end, consider Volterra filters, which consist of kernels of orders o,o{1,,O}Image, with maximum order O. Each kernel consists of diagonals with maximum relative time lags RoImage in the branch nonlinearities of the oth kernel (RoImage is also known as diagonal radius) and has linear subsystems of length LoImage in kernel o. As O, RoImage and LoImage determine the model structure, they will be referred to as the structure parameters of the Volterra model. These structure parameters determine which of the filter coefficients of a Volterra filter are modeled in practice.

As proposed in [43], an identified model can automatically be trimmed by an additional adaptive structure, where a Volterra kernel is clustered into disjoint groups of diagonals and each group's output signal yg(n)Image, where g is the group index, is replaced by a convex combination of the group's output yg(n)Image with 0 (the result of filtering with an all-zero kernel region). Thereby, irrelevant diagonals, which are dominated by noise and would increase the modeling error, can be trimmed for the final computation of the model output. Thereby, the modeling accuracy is increased at the expense of a slight increase in computational complexity. In this case, the mixing variables of the convex combinations take the role of additional structure parameters. In a similar way, the FLAFs in [48] disable nonlinear model parts as well. Another related approach is the so-called EVOLVE concept [76], which approaches the estimation of a Volterra filter structure in a bottom-up way by starting with a small filter and successively growing the model until the optimal size is reached. Filters which are evolutionary w.r.t. a structure parameter dimension S{O,Ro,Lo}Image perform a convex combination of a filter with a small model size (a low structure parameter value S1Image) and of a filter with a larger model size (a larger structure parameter value S2>S1Image). In the case of a clear preference of one of both models, both models are grown or shrunk into the preferred direction. Note that, for reasons of efficiency, the larger model can also be “virtualized” by simply augmenting the small model with additional filter coefficients [76,77]. As described in [76], an evolution in RoImage and LoImage and bypassing kernel orders is possible by cascading multiple convex combinations, each of which decides for the activation of another region of filter taps of the Volterra filter. This concept can, of course, be applied to CGMs with other branch nonlinearities as well: Legendre-based CGMs (cf., [45,46] or Section 4.5.2), where ith powers in the Volterra filters' branch nonlinearities are replaced by ith-order Legendre polynomials, have the same set of structure parameters {O,Ro,Lo}Image and can be grown the same way as Volterra filters in an EVOLVE fashion. Also, the model trimming from [43] can directly be applied to such Legendre-based CGMs. As opposed to [43], EVOLVE [76] starts with a low-complexity model and successively grows the model if required. Thereby, not only gradient noise in irrelevant filter coefficients is prevented, but also the computational complexity is only as high as the complexity of the optimum filter structure found by the algorithm, plus some overhead for the virtualized models of increased size.

Like in SA filtering, efficient EVOLVE implementations can employ a block partitioning suitable for PB-FNLMS adaptation. This allows one to grow and shrink the memory lengths of the diagonals in steps of M taps by adding or removing entire length-M partitions. However, the efficiency and accuracy are achieved differently from SA filtering. SA filtering estimates the nonlinear characteristics of the unknown system after decomposing the system into a nonlinear system with short nonzero temporal support and the estimated nonlinear characteristics are then extrapolated efficiently using CMs with parametric preprocessors. In EVOLVE, such a model decomposition and extrapolation does not take place, so SA filtering locally (at particular filter coefficients) observes the nonlinearity and extrapolates it to a global system representation, assuming a fixed model structure (CM structure). As opposed to this, EVOLVE locally grows the model (at particular diagonals in a particular Volterra kernel), aiming for the best-performing global model structure and parametrization as final result. Yet, the SA and the EVOLVE approach do not contradict each other but can jointly be implemented with the same block partitioning scheme. The resulting evolutionary SA filters may, in the future, dynamically grow their structure (e.g., number of partitions P and the number of branches B) while extrapolating nonlinearities with a preprocessor system, whenever the computational complexity restricts additional partitions or a direct estimation of the particular CGM partitions is disadvantageous due to adaptation noise.

4.7 Summary

In this chapter, recent advances in LIP nonlinear filters have been presented with the focus on the concept of Significance-Aware (SA) filtering and nonlinear AEC as a challenging application in the background. After a unified description of nonlinear models as variants of CGMs and LIP CMs, where the output is a bilinear function of the model parameters, the parameter estimation methods for both structures have been revisited, showing that some of the recent developments can be described as FX algorithms. This lead to the proposal of a novel, highly efficient FX algorithm for the adaptation of CMs with parametric nonlinear preprocessors. Afterwards, the SA-filtering concept has been described in two variants: the SSA-CM and the PSA-CGM, which gain efficiency by equalizing parts of or partly compensating the unknown system before estimating the system's nonlinearity. Additionally, a novel PSA FX algorithm has been introduced as well. The algorithms have been compared in AEC experiments with recordings from three different physical devices, which have shown the high efficacy of SA filtering and its potential to save computational complexity without sacrificing modeling accuracy. Finally, model structure estimation based on convex filter combinations has been summarized and potential synergies with SA filters in a joint partitioned-block signal processing framework have been pointed out.

References

[1] M.J. Korenberg, I.W. Hunter, The identification of nonlinear biological systems: Volterra kernel approaches, Annals of Biomedical Engineering March 1996;24(2):250–268.

[2] Y. Chen, I.W. Hunter, Nonlinear stochastic system identification of skin using Volterra kernels, Annals of Biomedical Engineering April 2013;41(4):847–862.

[3] S.A. Billings, Identification of nonlinear systems—a survey, IEEE Proc. Control Theory and Applications November 1980;127(6):272–285.

[4] S. Benedetto, E. Biglieri, R. Daffara, Modeling and performance evaluation of nonlinear satellite links—a Volterra series approach, IEEE Transactions on Aerospace and Electronic Systems July 1979;AES-15(4):494–507.

[5] K.V. Peddanarappagari, M. Brandt-Pearce, Volterra series transfer function of single-mode fibers, Journal of Lightwave Technology December 1997;15(12):2232–2241.

[6] G.M. Raz, B.D. van Veen, Baseband Volterra filters for implementing carrier based nonlinearities, IEEE Transactions on Signal Processing January 1998;46(1):103–114.

[7] W.A. Frank, An efficient approximation to the quadratic Volterra filter and its application in real-time loudspeaker linearization, Signal Processing July 1995;45(1):97–113.

[8] S. Kinoshita, Y. Kajikawa, Y. Nomura, Volterra filters using multirate signal processing and their application to loudspeaker systems, IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP, vol. 6. Salt Lake City, USA. May 2001:3497–3500.

[9] W. Frank, R. Reger, U. Appel, Loudspeaker nonlinearities—analysis and compensation, Asilomar Conference on Signals, Systems, and Computers, vol. 2. Pacific Grove, California, USA. October 1992:756–760.

[10] A. Stenger, L. Trautmann, R. Rabenstein, Nonlinear acoustic echo cancellation with 2nd order adaptive Volterra filters, IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP, vol. 2. Phoenix, Arizona, USA. March 1999:877–880.

[11] A. Fermo, A. Carini, G. Sicuranza, Analysis of different low complexity nonlinear filters for acoustic echo cancellation, Proceedings of the First International Workshop on Image and Signal Processing and Analysis. IWISPA, Pula, Croatia. June 2000:261–266.

[12] S. Kinoshita, Y. Kajikawa, Y. Nomura, Adaptive Volterra filters using multirate signal processing and their application to identification of loudspeaker systems, Electronics and Communications in Japan (Part III: Fundamental Electronic Science) July 2004;87(7):45–54.

[13] F. Küch, W. Kellermann, Partitioned block frequency-domain adaptive second-order Volterra filter, IEEE Transactions on Signal Processing February 2005;53(2):564–575.

[14] L.A. Azpicueta-Ruiz, M. Zeller, J. Arenas-García, W. Kellermann, Novel schemes for nonlinear acoustic echo cancellation based on filter combinations, IEEE International Conference on Acoustics, Speech, and Signal Processing. ICASSP, Taipei, Taiwan. April 2009:193–196.

[15] L.A. Azpicueta-Ruiz, M. Zeller, A. Figueiras-Vidal, J. Arenas-García, W. Kellermann, Adaptive combination of Volterra kernels and its application to nonlinear acoustic echo cancellation, IEEE Transactions on Audio, Speech, and Language Processing January 2011;19(1):97–110.

[16] M. Zeller, L.A. Azpicueta-Ruiz, J. Arenas-García, W. Kellermann, Adaptive Volterra filters with evolutionary quadratic kernels using a combination scheme for memory control, IEEE Transactions on Signal Processing April 2011;59(4):1449–1464.

[17] H. Buchner, J. Benesty, T. Gänsler, W. Kellermann, Robust extended multidelay filter and double-talk detector for acoustic echo cancellation, IEEE Transactions on Audio, Speech, and Language Processing September 2006;14(5):1633–1644.

[18] K. Shi, X. Ma, G. Zhou, A mutual information based double-talk detector for nonlinear systems, Annual Conference on Information Sciences and Systems. CISS. March 2008:356–360.

[19] V.J. Mathews, G.L. Sicuranza, Polynomial Signal Processing. New York (NY), USA: Wiley; May 2000.

[20] K. Shi, G. Zhou, M. Viberg, Hammerstein system linearization with application to nonlinear acoustic echo cancellation, Digital Signal Processing Workshop, 12th—Signal Processing Education Workshop, 4th. September 2006:183–186.

[21] K. Shi, G. Zhou, M. Viberg, Compensation for nonlinearity in a Hammerstein system using the coherence function with application to nonlinear acoustic echo cancellation, IEEE Transactions on Signal Processing December 2007;55(12):5853–5858.

[22] A. Birkett, R. Goubran, Limitations of handsfree acoustic echo cancellers due to nonlinear loudspeaker distortion and enclosure vibration effects, IEEE ASSP Workshop on Applications of Signal Processing to Audio and Acoustics. WASPAA, New Paltz, NY, USA. October 1995:103–106.

[23] A. Stenger, R. Rabenstein, An acoustic echo canceller with compensation of nonlinearities, European Signal Processing Conference. EUSIPCO, Island of Rhodes, Greece. September 1998:969–972.

[24] A. Stenger, W. Kellermann, Adaptation of a memoryless preprocessor for nonlinear acoustic echo cancelling, Signal Processing September 2000;80(9):1747–1760.

[25] A. Stenger, Kompensation akustischer Echos unter Einfluß von nichtlinearen Audiokomponenten. Berichte aus der Kommunikations- und Informationstechnik. Aachen: Shaker Verlag; April 2001;vol. 23 in German.

[26] M. Scarpiniti, D. Comminiello, R. Parisi, A. Uncini, Comparison of Hammerstein and Wiener systems for nonlinear acoustic echo cancelers in reverberant environments, 17th International Conference on Digital Signal Processing. DSP. July 2011:1–6.

[27] S. Shimauchi, Y. Haneda, Nonlinear acoustic echo cancellation based on piecewise linear approximation with amplitude threshold decomposition, International Workshop on Acoustic Signal Enhancement. IWAENC, Aachen, Germany. September 2012.

[28] C. Hofmann, C. Huemmer, W. Kellermann, Significance-aware Hammerstein group models for nonlinear acoustic echo cancellation, IEEE International Conference on Acoustics, Speech, and Signal Processing. ICASSP, Florence, Italy. May 2014:5934–5938.

[29] S. Van Vaerenbergh, L.A. Azpicueta-Ruiz, Kernel-based identification of Hammerstein systems for nonlinear acoustic echo-cancellation, IEEE International Conference on Acoustics, Speech, and Signal Processing. ICASSP, Florence, Italy. May 2014:3739–3743.

[30] C. Huemmer, C. Hofmann, R. Maas, W. Kellermann, The significance-aware EPFES to estimate a memoryless preprocessor for nonlinear acoustic echo cancellation, IEEE Global Conference on Signal Information Processing. GlobalSIP, Atlanta, GA. December 2014:557–561.

[31] C. Hofmann, M. Guenther, C. Huemmer, W. Kellermann, Efficient nonlinear acoustic echo cancellation by partitioned-block significance-aware Hammerstein group models, European Signal Processing Conference. EUSIPCO, Budapest, Hungary. August 2016:1783–1787.

[32] C. Hofmann, C. Huemmer, M. Guenther, W. Kellermann, Significance-aware filtering for nonlinear acoustic echo cancellation, EURASIP Journal on Advances in Signal Processing 2016;2016(1):113.

[33] F. Küch, W. Kellermann, Orthogonalized power filters for nonlinear acoustic echo cancellation, Signal Processing June 2006;86(6):1168–1181.

[34] D. Comminiello, M. Scarpiniti, R. Parisi, A. Uncini, A functional link based nonlinear echo canceller exploiting sparsity, International Workshop on Acoustic Echo and Noise Control. IWAENC, Tel Aviv, Israel. August 2010.

[35] D. Comminiello, L.A. Azpicueta-Ruiz, M. Scarpiniti, A. Uncini, J. Arenas-García, Functional link based architectures for nonlinear acoustic echo cancellation, Joint Workshop on Hands-Free Speech Communication and Microphone Arrays. HSCMA, Edinburgh, UK. May 2011:180–184.

[36] M. Mossi, C. Yemdji, N. Evans, C. Beaugeant, Non-linear acoustic echo cancellation using online loudspeaker linearization, IEEE Workshop on Applications of Signal Processing to Audio and Acoustics. WASPAA, New Paltz, NY, USA. October 2011:97–100.

[37] M.I. Mossi, C. Yemdji, N. Evans, C. Beaugeant, P. Degry, Robust and low-cost cascaded non-linear acoustic echo cancellation, IEEE International Conference on Acoustics, Speech, and Signal Processing. ICASSP, Prague, Czech Republic. IEEE; May 2011:89–92.

[38] F. Chang, R. Luus, A noniterative method for identification using Hammerstein model, IEEE Transactions on Automatic Control October 1971;16(5):464–468.

[39] P. Gallman, An iterative method for the identification of nonlinear systems using a Uryson model, IEEE Transactions on Automatic Control December 1975;20(6):771–775.

[40] F. Küch, A. Mitnacht, W. Kellermann, Nonlinear acoustic echo cancellation using adaptive orthogonalized power filters, IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP, vol. 3. Philadelphia, PA, USA. March 2005:105–108.

[41] M.I. Mossi, C. Yemdji, N. Evans, C. Beaugeant, A Cascaded Non-linear Acoustic Echo Canceller Combining Power Filtering and Clipping Compensation. [Technical Report RR-11-258, EURECOM Research Report] August 2011.

[42] E. Thomas, Some considerations on the application of the Volterra representation of nonlinear networks to adaptive echo cancellers, The Bell System Technical Journal October 1971;50(8):2797–2805.

[43] L.A. Azpicueta-Ruiz, M. Zeller, A. Figueiras-Vidal, W. Kellermann, J. Arenas-García, Enhanced adaptive Volterra filtering by automatic attenuation of memory regions and its application to acoustic echo cancellation, IEEE Transactions on Signal Processing June 2013;61(11):2745–2750.

[44] A. Carini, G. Sicuranza, Even mirror Fourier nonlinear filters, IEEE International Conference on Acoustics, Speech, and Signal Processing. ICASSP, Vancouver, BC, Canada. May 2013:5608–5612.

[45] A. Carini, S. Cecchi, M. Gasparini, G. Sicuranza, Introducing Legendre nonlinear filters, IEEE International Conference on Acoustics, Speech, and Signal Processing. ICASSP, Florence, Italy. May 2014:7939–7943.

[46] A. Carini, S. Cecchi, L. Romoli, G.L. Sicuranza, Perfect periodic sequences for Legendre nonlinear filters, European Signal Processing Conference. EUSIPCO. September 2014:2400–2404.

[47] A. Carini, G.L. Sicuranza, A study about Chebyshev nonlinear filters, Signal Processing May 2016;122:24–32.

[48] D. Comminiello, M. Scarpiniti, L.A. Azpicueta-Ruiz, J. Arenas-García, A. Uncini, Functional link adaptive filters for nonlinear acoustic echo cancellation, IEEE Transactions on Audio, Speech, and Language Processing July 2013;21(7):1502–1512.

[49] D. Comminiello, M. Scarpiniti, L.A. Azpicueta-Ruiz, J. Arenas-García, A. Uncini, Nonlinear acoustic echo cancellation based on sparse functional link representations, IEEE/ACM Transactions on Audio, Speech, and Language Processing July 2014;22(7):1172–1183.

[50] D. Comminiello, M. Scarpiniti, L.A. Azpicueta-Ruiz, J. Arenas-García, A. Uncini, A nonlinear architecture involving a combination of proportionate functional link adaptive filters, European Signal Processing Conference. EUSIPCO. August 2015:2869–2873.

[51] K. Narendra, P. Gallman, An iterative method for the identification of nonlinear systems using a Hammerstein model, IEEE Transactions on Automatic Control July 1966;11(3):546–550.

[52] E. Eskinat, S.H. Johnson, W.L. Luyben, Use of Hammerstein models in identification of nonlinear systems, American Institute of Chemical Engineers (AIChE) Journal February 1991;37(2):255–268.

[53] G. Sicuranza, A. Carini, On the accuracy of generalized Hammerstein models for nonlinear active noise control, IEEE Instrumentation and Measurement Technology Conference. IMTC. April 2006:1411–1416.

[54] E.-W. Bai, K. Li, Convergence of the iterative algorithm for a general Hammerstein system identification, Automatica November 2010;46(11):1891–1896.

[55] S. Malik, G. Enzner, Fourier expansion of Hammerstein models for nonlinear acoustic system identification, IEEE International Conference on Acoustics, Speech, and Signal Processing. ICASSP, Prague, Czech Republic. May 2011:85–88.

[56] C. Huemmer, C. Hofmann, R. Maas, A. Schwarz, W. Kellermann, The elitist particle filter based on evolutionary strategies as novel approach for nonlinear acoustic echo cancellation, IEEE International Conference on Acoustics, Speech, and Signal Processing. ICASSP, Florence, Italy. May 2014:1315–1319.

[57] S. Haykin, Adaptive Filter Theory. 4th edition Upper Saddle River (NJ), USA: Prentice Hall; 2002.

[58] J.-S. Soo, K. Pang, Multidelay block frequency domain adaptive filter, IEEE Transactions on Acoustics, Speech and Signal Processing February 1990;38(2):373–376.

[59] H. Buchner, J. Benesty, T. Gänsler, W. Kellermann, An outlier-robust extended multidelay filter with application to acoustic echo cancellation, International Workshop on Acoustic Echo and Noise Control. IWAENC, Kyoto, Japan. September 2003:19–22.

[60] M. Zeller, W. Kellermann, Self-configuring system identification via evolutionary frequency-domain adaptive filters, International Workshop on Acoustic Echo and Noise Control. IWAENC, Tel Aviv, Israel. August 2010.

[61] F. Küch, E. Mabande, G. Enzner, State-space architecture of the partitioned-block-based acoustic echo controller, IEEE International Conference on Acoustics, Speech, and Signal Processing. ICASSP, Florence, Italy. May 2014:1295–1299.

[62] G. Enzner, P. Vary, A soft-partitioned frequency-domain adaptive filter for acoustic echo cancellation, IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP, vol. 5. Hong Kong. April 2003:V-393–V-396 (cancelled due to the SARS epidemic).

[63] P.C.W. Sommen, Adaptive Filtering Methods. [PhD thesis] Technische Universiteit Eindhoven; 1992.

[64] S.M. Kuo, D.R. Morgan, Active Noise Control Systems: Algorithms and DSP Implementations. Wiley; February 1996.

[65] B. Widrow, D. Shur, S. Shaffer, On adaptive inverse control, Asilomar Conference on Circuits, Systems, and Computers. Pacific Grove, California, USA. September 1981:185–189.

[66] Y.A. Huang, J. Skoglund, A. Luebs, Practically efficient nonlinear acoustic echo cancellers using cascaded block RLS and FLMS adaptive filters, IEEE International Conference on Acoustics, Speech, and Signal Processing. ICASSP, New Orleans, LA, USA. March 2017:596–600.

[67] E.L.O. Batista, O.J. Tobias, R. Seara, A sparse-interpolated scheme for implementing adaptive Volterra filters, IEEE Transactions on Signal Processing April 2010;58(4):2022–2035.

[68] G.L. Sicuranza, A. Carini, Nonlinear active noise control, European Signal Processing Conference. EUSIPCO, Vienna, Austria. September 2004:1887–1890.

[69] L. Tronchin, The emulation of nonlinear time-invariant audio systems with memory by means of Volterra series, Journal of the Audio Engineering Society December 2012;60(12):984–996.

[70] L. Tronchin, V.L. Coli, Further investigations in the emulation of nonlinear systems with Volterra series, Journal of the Audio Engineering Society September 2015;63(9):671–683.

[71] A. Krzyżak, J.Z. Sasiadek, S. Ulrich, Nonparametric identification of robot flexible joint space manipulator, 17th International Conference on Methods Models in Automation Robotics. MMAR, Miȩdzyzdroje, Poland. August 2012:172–177.

[72] J. Lee, B. Bagheri, H.-A. Kao, A cyber-physical systems architecture for Industry 4.0-based manufacturing systems, Manufacturing Letters January 2015;3(Supplement C):18–23.

[73] G. Enzner, T.C. Kranemann, P. Thüne, Evaluation of estimated Hammerstein models via normalized projection misalignment of linear and nonlinear subsystems, IEEE International Conference on Acoustics, Speech, and Signal Processing. ICASSP. March 2016:4234–4238.

[74] C. Breining, P. Dreiseitel, E. Hänsler, A. Mader, B. Nitsch, H. Puder, T. Schertler, G. Schmidt, J. Tilp, Acoustic echo control. An application of very-high-order adaptive filters, IEEE Signal Processing Magazine July 1999;16(4):42–69.

[75] J. Arenas-García, A. Figueiras-Vidal, A. Sayed, Mean-square performance of a convex combination of two adaptive filters, IEEE Transactions on Signal Processing March 2006;54(3):1078–1090.

[76] M. Zeller, Generalized Nonlinear System Identification Using Adaptive Volterra Filters with Evolutionary Kernels. München: Dr. Hut Verlag; 2013.

[77] M. Zeller, L.A. Azpicueta-Ruiz, W. Kellermann, Adaptive FIR filters with automatic length optimization by monitoring a normalized combination scheme, IEEE Workshop on Applications of Signal Processing to Audio and Acoustics. WASPAA, New Paltz, NY, USA. October 2009:149–152.


1  “This should not be confused with the bilinear filters in [19], which are recursive filters where the difference equation contains linear and bilinear terms in x(n)Image and y(n)Image.”

2  “Note that each filter tap in h(b)Image can alternatively be expressed as an additional preprocessor branch with an appropriately delayed version of fb{}Image as branch nonlinearity.”

3  “Note that [66] uses a different notation than this chapter and refers to an HM with a monome-based preprocessor as power filter.”

4  “Also note that not every bilinear cascade calls for FX adaptation, e.g., in the so-called sparse-interpolated Volterra filters of [67], linear interpolation filters are prefilters to subsampled Volterra kernels to interpolate the latter. As the interpolation filters are time-invariant, they do not need to be adapted and consequently do not require an FX adaptation.”

5  “As opposed to the equalization-based SA-HM [32], the equalizer matches its output with xpp(nNeq)Image instead of the input signal x(n)Image itself. Thereby, the error signal eeq(n)Image used for adapting the equalizer can actually approach zero when identifying the preprocessor correctly and consequently leads to less gradient noise in the adaptation process.”

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
18.218.156.35