Chapter 3

Spline Adaptive Filters

Theory and Applications

Michele ScarpinitiDanilo ComminielloRaffaele ParisiAurelio Uncini    Sapienza University of Rome, Rome, Italy

Abstract

In this chapter we propose a unifying overview of a recent class of nonlinear adaptive filters, known as Spline Adaptive Filter (SAF). This family of nonlinear adaptive filters comes from the general class of block-wise architectures consisting of a cascade of a suitable number of linear blocks and flexible memoryless nonlinear functions. In particular, the nonlinear function involved in the adaptation process is based on a spline function whose shape can be modified during the learning. Specifically, B-splines and Catmull–Rom splines are used, because they allow one to impose simple constraints on control parameters. The spline control points are adaptively changed using gradient-based techniques. In addition, in this chapter we show some of the most meaningful theoretical results in terms of upper bounds on the choice of the step sizes and excess mean square error values.

The SAF family of nonlinear adaptive filters can be successfully applied to several real-world applications, such as the identification of nonlinear systems, adaptive echo cancelers, adaptive noise control, nonlinear prediction and some other learning algorithms. Some experimental results are also presented to demonstrate the effectiveness of the proposed method.

Keywords

Nonlinear adaptive filters; Spline adaptive filters; Nonlinear cascade adaptive filters; Least mean square; Wiener systems; Hammerstein systems; Sandwich models

Chapter Points

  • •  We describe a nonlinear filtering approach based on spline nonlinear functions.
  • •  The proposed approach can be successfully used for the identification of nonlinear systems.
  • •  The proposed approach outperforms other approaches based on Volterra filters, both in speed of convergence and in performance.
  • •  We provide upper bounds on the choice of learning rates and the theoretical EMSE estimates.

Acknowledgements

This chapter has been partially supported by the PRIN2015 project no. 2015YPXH4W_004: “A green adaptive Fog computing and networking architecture (GAUChO)”, funded by the Italian MIUR, and by the project: “Vehicular Fog for energy-efficient QoS mining and dissemination of multimedia Big Data streams (V-FOG),” funded by Sapienza University of Rome.

3.1 Introduction

The theory of linear adaptive filtering (LAF) is well established in the literature [13]. In addition, the use of LAF methods for the solution of real-world problems is extensive and represents a central paradigm for many strategic applications [4]. The basic discrete-time linear adaptive filter model is made of a simple linear combiner whose length is determined by following some proper criterion and whose weights are adapted by choosing the most convenient optimization strategy [1].

In contrast, in many practical situations, the nature of the process of interest is often affected by any nonlinearity and thus requires the use of nonlinear adaptive filter (NAF) models. In the ideal case, if the nature of the nonlinearities is known a priori, it is possible to determine the NAF structure, in analogy with the mathematical model of nonlinearities. In this case the NAF is referred to as white-box (or structural) model and the process of adaptation consists of the estimation of the free model parameters (e.g., the coefficients of the nonlinear difference equation) [5].

On the contrary, if any a priori knowledge about the nature of nonlinearity is not available, it is preferred to use a black-box approach (functional model) [6,7]. In these cases, a unique framework for the solution does not exist and the existence of the solution cannot be guaranteed.

One of the best-known functional models that provide a good and general solution is based on the Volterra series [8]. Unfortunately, due to the large number of free parameters required, the Volterra adaptive filter (VAF) is generally employed only in situations of mild nonlinearity [8,9]. Neural networks (NNs) [10,11] represent another flexible tool to realize black-box NAFs, but they generally require a high computational cost and are not always feasible in online applications.

In the literature on nonlinear filtering, one of the most used structures is the so-called block-oriented representation, in which linear time-invariant (LTI) models are connected with memoryless nonlinear functions. The basic classes of block-oriented nonlinear systems are represented by the Wiener and Hammerstein models [1214] and all those architectures originated by the connection of these two classes according to different topologies (i.e., parallel, cascade, feedback, etc. [15]). More specifically, the Wiener model consists of a cascade of an LTI filter followed by a static nonlinear function and sometimes it is known as linear–nonlinear (LN) model [1619]. Conversely, the Hammerstein model consists of a cascade connection of a static nonlinear function followed by an LTI filter [2022] and is sometimes indicated as nonlinear–linear (NL) model.

In a general nonlinear problem, with no a priori information available, the best adaptive model (Hammerstein or Wiener) is that most similar to the one to be identified. This fact implies that Wiener and Hammerstein models could be too restrictive in practical applications. Hence, other cascade models, such as the linear–nonlinear–linear (LNL) or the nonlinear–linear–nonlinear (NLN) models, were also introduced [12,23,24]. These cascade models, here called sandwich models, constitute a relatively simple but important class of nonlinear models that can approximate a wide class of nonlinear systems [25].

These models assume a key role in practical applications, since they are often related to specialized classes of nonlinear systems. As a matter of fact, these architectures can be successfully used in several applications in all fields of engineering [26], from circuit and system theory to modeling nonlinear characteristics of electronic devices and resistive networks [2729], to signal processing [19,30], biomedical data analysis [31], digital communications [32,33] and even to hydraulics [34] and chemistry [35].

A crucial point in adopting a block-oriented model is the choice of the specific nonlinear function to be adopted. Obviously, several choices for this nonlinearity are possible, namely some invertible squashing functions, piece-wise linear functions [36,37,27], polynomial nonlinearities [38], splines [39] and more.

Recently a novel approach for the identification of Wiener, Hammerstein and sandwich models based on Spline Adaptive Filters (SAFs) has been proposed [30,4044]. Among the advantages of these solutions, in order to yield great representative potential, high performance, low computational cost and easy implementability in dedicated hardware, all the nonlinearities involved in such systems are flexible and adapted by means of a fixed-length look-up table (LUT) that collects a suitable number of samples of the nonlinearity. Moreover, in order to have a continuous and derivable smooth function, the LUT samples are interpolated by a low-order uniform spline, like done in [40] for the Wiener SAF (WSAF), in [41] for the Hammerstein SAF (HSAF) and in [43] for the two sandwich SAFs (S1SAF and S2SAF).

SAFs have been successfully adopted in the literature in different scenarios, such as nonlinear echo cancellation [30] and active noise control [4547], where signals are distorted by the low-cost hardware equipments. The standard LMS adaptation scheme has also been extended to the normalized version in [48] and to the maximum correntropy criterion [49]. The latter approach is particularly efficient in the case when data are corrupted by impulsive noise or contain large outliers. Finally, the SAF approach has been recently extended to the diffusion case [50], involving a network of distributed sensors.

In this chapter we aim at providing a deep and unifying overview of SAF models. After the description of the basic architectures and related learning algorithms, the main theoretical results on the choice of the step-size and the steady-state mean square error (MSE) are also provided. Finally some experimental results show the effectiveness of this family of nonlinear adaptive filters to solve real-world applications.

The rest of the chapter is organized as follows. Section 3.2 gives some background of spline interpolation. Section 3.3 provides a complete description of SAF models including the computational cost. Section 3.4 is devoted to the introduction of convergence results, while Section 3.5 provides some applicative scenarios and numerical results. Finally, Section 3.6 concludes the chapter.

3.1.1 Notation

In this chapter matrices are represented by boldface capital letters, i.e., ARM×NImage, while Tr()Image denotes the trace of a matrix. All vectors are column vectors, denoted by boldface lowercase letters, like wRM×1=[w[0],w[1],,w[M1]]TImage, where w[i]Image denotes the ith individual entry of w. In the recursive algorithm definition, a discrete-time subscript index n is added. For example, the weight vector, calculated according to some law, is written as wn+1=wn+ΔwnImage. In the case of signal regression, vectors are indicated as xnRM×1=[x[n],x[n1],,x[nM+1]]TImage. Moreover, regarding the interpolated nonlinearity, the LUT control points are denoted qnRQ×1=[q1,q2,,qQ]TImage, where qjImage is the jth control point, while the ith span is denoted by qi,nR4×1=[qi,qi+1,qi+2,qi+3]TImage. Finally, the Euclidean norm and the transpose of a vector are denoted by Image and ()TImage, respectively.

3.2 Foundation of Spline Interpolation

Historically derived from elastic rulers that cross a certain number of points, called knots, and used in the past for technical drawing, splines are differentiable, up to a given order, polynomial curves for interpolation or approximation of a succession of Q+1Image knots inside a convex hull. The knots Qi=[qx,i,qy,i]TImage, for i=0,1,,QImage, represent a tabulated function on the plane x-y [51,39,52] and, in order to construct an appropriate function, they are also constrained as

q x , 0 < q x , 1 < < q x , Q .

Image (3.1)

Literature on splines is vast. Without loss of generality, in this brief review we consider the univariate basis or blending functions developed by Schoenberg, who introduced the name B-splines, deriving them from Bézier splines [51].

Let u[qx,i,qx,i+1]Image be the abscissa value between two consecutive knots. The spline curve is defined as an affine combination of some knots of a suitable P degree spline basis function NiP(u)Image with minimal support and certain continuity properties [52]:

φ ( u ) = i = 0 Q P 1 Q i N i P ( u ) , u [ q x , n , q x , Q P ] .

Image (3.2)

Given Q+1Image knots, the univariate spline basis functions NiP(u)Image can be defined by the Cox–de Boor recursion [39]

N i 0 ( u ) = { 1 , q x , i u < q x , i + 1 , 0 , otherwise, N i P ( u ) = u q x , i q x , i + P q x , i N i P 1 ( u ) + q x , i + P + 1 u q x , i + P + 1 q x , i + 1 N i + 1 P 1 ( u ) ,

Image (3.3)

where i=0,1,,QP1Image. The blending functions NiP(u)Image are polynomials of degree P with continuity of the first P1Image derivatives. For example, for P=1Image, we have

N i 1 ( u ) = { u q x , i q x , i + 1 q x , i , q x , i u < q x , i + 1 , q x , i + 2 u q x , i + 2 q x , i + 1 , q x , i + 1 u < q x , i + 2 , 0 , otherwise,

Image

which results in a classical linear interpolation scheme, so Ni0(u)Image is a rectangular function, Ni1(u)Image is a linear double-ramp function, Ni2(u)Image is a quadratic double-ramp function. The generic NiP(u)Image can be evaluated as a sequential convolution of P rectangular pulse functions NiP=Ni0Ni0Image.

Hence, the curve (3.2) can be written as

φ ( u ) = φ i ( u ) , i = 0 , , Q  and  u [ 0 , 1 ) ,

Image (3.4)

where φi(u)Image is a P degree local polynomial called curve span, entirely defined from P+1Image knots, that are sometimes called control points. Moreover, scaling or translating the knot vector does not modify the basis functions, that can be fixed a priori for all control points.

From (3.4), there must be a unique mapping that allows us to calculate the local parameter u, as well as the proper curve span index i, from the global abscissa parameter (that in our case is the input sequence to the nonlinearity). In this way, we can represent any point lying on the curve φ(u)Image as a point belonging to the single φi(u)Image curve span.

3.2.1 Uniform Cubic Spline

The recursive Eq. (3.3) can be entirely pre-calculated with considerable computational saving, by imposing a uniform distribution knot-vector, i.e., constant sampling step of the abscissa Δx=qx,iqx,i+1Image [39]. This implies that the local polynomial φi(u)Image in (3.4) can be written in a very simple form. A similar approach was successfully used in multilayer feed-forward neural networks [53].

Regarding the choice of the optimal degree P of a spline basis, a good compromise between flexibility and approximation behavior is represented by uniform cubic spline nonlinearities [54], as they provide good approximation capabilities avoiding the problem of over-fitting.

In the case of our interest where P=3Image, the precalculation of the recursion (3.3) provides [39,54,53,40]

N i 3 ( u ) = { 1 6 [ ( u q i ) 3 + 3 ( u q i ) 2 3 ( u q i ) + 1 ] , q x , i u < q x , i + 1 , 1 6 [ 3 ( u q i + 1 ) 3 6 ( u q i + 1 ) 2 + 4 , ] , q x , i + 1 u < q x , i + 2 , 1 6 [ 3 ( u q i + 2 ) 3 + 3 ( u q i + 2 ) 2 + 3 ( u q i + 2 ) + 1 ] , q x , i + 2 u < q x , i + 3 , 1 6 ( u q i + 3 ) 3 , q x , i + 3 u < q x , i + 4 ,

Image

which can be reformulated in a matrix form as

φ i ( u ) = [ u 3 u 2 u 1 ] C [ q i q i + 1 q i + 2 q i + 3 ] ,

Image (3.5)

for i=0,1,,QImage and u[0,1)Image and where the basis matrix C is defined as

C = 1 6 [ 1 3 3 1 3 6 3 0 3 0 3 0 1 4 1 0 ] .

Image (3.6)

Imposing different constraints to the approximation relationship, several spline bases with different properties can be evaluated in a similar manner. An example of such a basis is the Catmull–Rom (CR) spline [55], very important in many applications. The interpolation scheme is the same of (3.5), while the CR-spline basis matrix C has the form

C = 1 2 [ 1 3 3 1 2 5 4 1 1 0 1 0 0 2 0 0 ] .

Image (3.7)

Note that CR-splines, initially developed for computer graphics purposes, represent a local interpolating scheme. The CR-spline, in fact, specifies a curve that passes through all of the control points, a feature which is not necessarily true for other spline methodologies. Overall, the CR-spline results in a more local approximation with respect to the B-spline.

For both the CR- and B-basis the curve has a continuous first derivative. For the B-spline the second derivative is also a continuous function, while for the CR spline the second derivative is not continuous at the knots. However, there is a regularization property common to most of the polynomial spline basis set (CR- and B-spline included) called variation diminishing property [39], which ensures the absence of unwanted oscillations of the curve between two consecutive control points, as well as the exact representation of linear segments.

Note also that the entries of the matrix C satisfy an important general property: the sum of all the entries is equal to one. This property is a direct consequence of the fact that the spline basis functions, derived from Bézier curves, are a generalization of the Bernstein polynomials and hence they represent a partition of unit (i.e., the sum of all basis function is one) [39,51].

In the next section, we present the main ideas on the SAFs and the derivation of the basic SAF models.

3.3 Spline Adaptive Filters

A Spline Adaptive Filter (SAF) consists of the cascade of a suitable number of linear filters and memoryless nonlinear functions obtained by the spline interpolation scheme described in Section 3.2.

In order to compute the nonlinear output y[n]=φ(x[n])Image, which is a spline interpolation of some adaptive control points contained in a LUT, we have to determine the explicit dependence between the input signal x[n]Image and y[n]Image. This can easily be done by considering that y[n]Image is a function of two local parameters unImage and i that depend on x[n]Image. From this onward we denote the local abscissa unImage with the subscript n in order to highlight the dependence on the input sample x[n]Image.

In the simple case of a uniformly spaced spline and a third-order (P=3Image) local polynomial interpolator (adopted in this work), each spline segment, denoted span, is controlled by four (P+1Image) control points, also known as knots, addressed by the span index i that points at the first element of the related four points contained in the LUT. The computation procedure for the determination of the span index i and the local parameters unImage can be expressed by the following equations [53]:

u n = x [ n ] Δ x x [ n ] Δ x , i = x [ n ] Δ x + Q 1 2 ,

Image (3.8)

where Δx=qi+1qiImage is the uniform space between knots, Image is the floor operator and Q is the fixed LUT length, i.e., the total number of control points. The second term on the right side of the second equation is an offset value needed to force i to be always nonnegative. Note that the index i depends on the time index n, i.e., inImage; for simplicity of notation in the following we will write simply iniImage.

Referring to [40] and to [41] and with reference to (3.5), the output of the nonlinearity can be evaluated as

y [ n ] = φ i ( u n ) = u n T Cq i , n ,

Image (3.9)

where the vector unImage is defined as unR4×1=[un3,un2,un,1]TImage, CR4×4Image is the precomputed matrix in (3.6) or (3.7), called spline basis matrix, and the vector qi,nImage contains the control points at the instant n. More specifically, qi,nImage is defined by qi,nR4×1=[qi,qi+1,qi+2,qi+3]TImage, where qkImage is the kth entry in the LUT.

It could be very important to compute the derivative of (3.9) with respect to its input. It is easily evaluated as

φ i ( u n ) = u ˙ n T Cq i , n ,

Image (3.10)

where u˙nR4×1=[3un2,2un,1,0]TImage.

The online learning algorithm can be derived by considering the cost function1 (CF) Jˆ(wn,qi,n)=E{e2[n]}Image. In order to derive the LMS algorithm, this CF is usually approximated by considering only the instantaneous error

J ( w n , q i , n ) = e 2 [ n ] .

Image (3.11)

The expression of the error signal e[n]Image in (3.11) depends on the particular architecture used.

In the following subsections we provide the description of the four basic SAF architectures.

3.3.1 Wiener SAF

The Wiener SAF (WSAF) architecture consists of the cascade of a linear filter, represented by the filter coefficient weight vector wnImage, and a memoryless nonlinear function implemented by a spline interpolation scheme whose knots are denoted by the vector qnImage. The Wiener architecture is shown in Fig. 3.1, where s[n]Image denotes the output of the linear filter and it is the input to the spline function.

Image
Figure 3.1 Block diagram of a WSAF.

With reference to Fig. 3.1, let us define the output error e[n]Image as

e [ n ] = d [ n ] y [ n ] = d [ n ] φ i ( u n ) ,

Image (3.12)

where d[n]Image is the reference signal, i.e., the output of the model to be identified. The online learning algorithm can be derived by the minimization of (3.11). We proceed by applying the stochastic gradient adaptation [40], obtaining the LMS iterative algorithm

w n + 1 = w n + μ w [ n ] e [ n ] φ i ( u ) x n ,

Image (3.13)

for the weights of the linear filter and

q i , n + 1 = q i , n + μ q [ n ] e [ n ] C T u n ,

Image (3.14)

for the spline control points. The parameters μw[n]Image and μq[n]Image represent the learning rates for the weights and for the control points, respectively, and, for simplicity, incorporate the other constant values.

3.3.2 Hammerstein SAF

The Hammerstein SAF (HSAF) architecture consists of the cascade of a memoryless nonlinear function implemented by a spline interpolation scheme whose knots are denoted by the vector qnImage and a linear filter, represented by the filter coefficient weight vector wnImage. The Hammerstein architecture is shown in Fig. 3.2, where s[n]Image denotes the output of the nonlinear function.

Image
Figure 3.2 Block diagram of an HSAF.

With reference to Fig. 3.2, we can define the output error e[n]Image as

e [ n ] = d [ n ] y [ n ] = d [ n ] w n T s n ,

Image (3.15)

where d[n]Image is the reference signal, sn=[s[n],s[n1],,s[nM+1]]TImage, and each sample s[nk]Image is evaluated by using (3.9).

As in the Wiener case, the learning algorithm is derived by minimizing the CF (3.11). We proceed by applying the stochastic gradient adaptation as done in [41], obtaining the LMS iterative algorithm

w n + 1 = w n + μ w [ n ] e [ n ] s n

Image (3.16)

for the weights of the linear filter and

q i , n + 1 = q i , n + μ q [ n ] e [ n ] C T U i , n w n

Image (3.17)

for the spline control points. In (3.17), Ui,nR4×M=[ui,n,ui,n1,,ui,nM+1]Image is a matrix which collects M past vectors ui,nkImage, where each vector assumes the value of unkImage if evaluated in the same span i of the current sample input x[n]Image or in an overlapped span; otherwise it is a zero vector.

As in the Wiener case, the learning rates μw[n]Image and μq[n]Image incorporate the other constant values.

3.3.3 Sandwich 1 SAF

The Sandwich 1 (S1SAF) architecture [43] consists of the cascade of a memoryless nonlinear function implemented by a spline interpolation scheme whose knots are denoted by the vector qn(1)Image, a linear filter, represented by the filter coefficient weight vector wnImage, and a second memoryless nonlinear function implemented by a spline interpolation scheme whose knots are denoted by the vector qn(2)Image. The S1 architecture is shown in Fig. 3.3, where s[n]Image denotes the output of the first nonlinear function and r[n]Image is the output of the linear filter.

Image
Figure 3.3 Block diagram of an S1SAF.

With reference to Fig. 3.3, we can define the output error e[n]Image as

e [ n ] = d [ n ] y [ n ] = d [ n ] φ ( 2 ) ( w n T s n ) ,

Image (3.18)

where d[n]Image is the reference signal. As in the previous cases, the learning algorithm is derived by minimizing the CF (3.11). We proceed by applying the stochastic gradient adaptation as done in [43], obtaining the LMS iterative algorithm

q i , n + 1 ( 2 ) = q i , n ( 2 ) + μ q ( 2 ) [ n ] e [ n ] C T u n ( 2 )

Image (3.19)

for the control points of the second nonlinearity,

w n + 1 = w n + μ w [ n ] e [ n ] φ i ( u n ( 2 ) ) s n

Image (3.20)

for the weights of the linear filter and

q i , n + 1 ( 1 ) = q i , n ( 1 ) + μ q ( 1 ) [ n ] e [ n ] φ i ( u n ( 2 ) ) C T U i , n ( 1 ) w n

Image (3.21)

for the control points of the first nonlinearity.

As in the previous cases, the learning rates μw[n]Image, μq(1)[n]Image and μq(2)[n]Image incorporate the other constant values.

3.3.4 Sandwich 2 SAF

The Sandwich 2 (S2SAF) architecture [43] consists of the cascade of a linear filter, represented by the filter coefficient weight vector wn(1)Image, a memoryless nonlinear function implemented by a spline interpolation scheme whose knots are denoted by the vector qnImage and a second linear filter, represented by the filter coefficient weight vector wn(2)Image. The S2 architecture is shown in Fig. 3.4, where s[n]Image denotes the output of the first linear filter and r[n]Image is the output of the nonlinear function.

Image
Figure 3.4 Block diagram of an S2SAF.

With reference to Fig. 3.4, we can define the output error e[n]Image as

e [ n ] = d [ n ] y [ n ] = d [ n ] w n ( 2 ) T φ ( w n ( 1 ) T x n ) ,

Image (3.22)

where again d[n]Image is the reference signal. As in the previous cases, the learning algorithm is derived by minimizing the CF (3.11). We proceed by applying the stochastic gradient adaptation as done in [43], obtaining the LMS iterative algorithm

w n + 1 ( 2 ) = w n ( 2 ) + μ w ( 2 ) [ n ] e [ n ] r n

Image (3.23)

for the weights of the second linear filter,

q i , n + 1 = q i , n + μ q [ n ] e [ n ] C T U i , n w n ( 2 )

Image (3.24)

for the control points of the nonlinearity and

w n + 1 ( 1 ) = w n ( 1 ) + μ w ( 1 ) [ n ] e [ n ] φ i ( u n ) X n w n ( 2 )

Image (3.25)

for the weights of the first linear filter.

As in the previous cases, the learning rates μw(1)[n]Image, μw(2)[n]Image and μq[n]Image incorporate the other constant values.

3.3.5 Other Architectures

It is possible to derive other kinds of block-oriented architectures by suitable combinations of the basic architectures introduced in the previous subsections, thus obtaining particular cascade and/or parallel systems that can be successfully used to identify specific types of nonlinear systems or by using an IIR linear filter instead of a FIR one, like done in [42]. A more general situation of particular interest is the case of a convex (and affine) combination of the former architectures, like the ones proposed in [56] for another important class of nonlinear adaptive filters. In this scenario, the whole combined filter is able to automatically detect the subsystem that can provide better results. The extension of SAF to a combined scenario will be postponed to future works.

In a similar fashion, the work [46] introduces a hybrid nonlinear spline filter, which is designed as a cascade of an adaptive spline function and a single-layer adaptive nonlinear network, devoted to the problem of removing noise in the acoustic domain. Another work worthy to consider is [57], where the HSAF architecture is generalized to a Functional Link network. The proposed work shows very interesting results also in terms of classification ability.

Another type of generalization of SAFs is based on a suitable choice of the learning schemes. In most works, SAFs are adapted by using a general LMS strategy. However, this situation cannot be the best one. An extension to the more general normalized LMS (NLMS) is provided in [48], while the extension to the correntropy criterion can be found in [49]. The latter strategy is particularly useful when data are corrupted by impulsive noise or contain large outliers.

Finally, [50] derives a diffused version of the WSAF (D-SAF) by applying ideas from the Diffusion Adaptation (DA) framework. DA algorithms allow a network of agents to collectively estimate a parameter vector, by jointly minimizing the sum of their local cost functions. Experimental evaluations show that the D-SAF is able to robustly learn the underlying nonlinear model, with a significant gain compared with a noncooperative solution [50]. A detailed description of D-SAFs can be found in Chapter 9.

3.3.6 Computational Cost

From a computational point of view each SAF architecture requires the evaluation of two or three processing blocks. For each block the computational cost of the forward phase has to be evaluated (i.e., the evaluation of the system output) and also that of the backward phase (i.e., the adaptation of the adaptive system). It is well known from the literature that the computational cost of the LMS algorithm is 2M+1Image multiplications plus 2M additions for each input sample [1], if M denotes the filter length.

In addition the cost of the spline evaluation must be considered. For each iteration only the ith span of the curve is modified by calculating the quantities unImage, i and the expressions unTCqi,nImage, u˙nTCqi,nImage and CTunImage. Note that the calculation of the quantity Cqi,nImage is executed during the output computation, as well as in the backward phase (the spline derivatives). However, most of the computations can be done through the reuse of past calculations (e.g., by explicating the polynomial and by use of the well-known Horner rule [58]). The cost for the spline output computation and its adaptation is 4KMImage multiplications plus 4KAImage additions, where KMImage and KAImage are constants (less than 16), depending on the implementation structure (see for example [53]). In any case, for high deep memory SAF, where the filter length is M>>4Image, the computational overhead, for the nonlinear function computation and its adaptation, can be neglected with respect to a simple linear filter. It is straightforward to note that the computational cost can be further reduced by using some smart trick, like choosing the sampling step Δx or the learning rates μj[n]Image as a power of two. In fact, these constraints transform the multiplications into shift operations.

The total computational cost, inclusive of the forward and backward computation, for the presented four SAF architectures is summarized in Table 3.1. In this table M denotes the linear filter length. In the case of S2SAF architecture, M1Image and M2Image are the length of the first and second linear filter, respectively.

Table 3.1

Computational cost of the SAF architectures in terms of additions and multiplications
Additions Multiplications
WSAF 2M + 48 2M + 67
HSAF 23M + 20 30M + 14
S1SAF 23M + 68 30M + 81
S2SAF 2M1M2 + M1 + 22M2 + 35 2M1M2 + 30M2 + 38

3.4 Convergence Properties

The convergence properties of the proposed learning algorithms could be evaluated in terms of their stability behavior, convergence speed, accuracy of the steady state and transient results and tracking ability as a function of their MSEs. For simplicity, in this chapter the attention is focused only on the optimal choice of the learning rates and the theoretical formulation of the excess MSE (EMSE) for the Wiener and Hammerstein architectures. Additional details and other useful results can be found in [40,41,43,44].

3.4.1 Bounds on the Learning Rates

A necessary but not sufficient condition for the algorithm convergence is the identification of a bound on the learning rate choice. These bounds on the learning rates can be derived by the Taylor series expansion of the error e[n+1]Image around the time index n, stopped at the first order. This series expansion involves the derivatives of the error with respect to the free parameters of the considered architecture.

By performing these evaluations for the four SAF architectures introduced in Section 3.3 and by using the previous notation, we can easily obtain

0 < μ w [ n ] φ i 2 ( u ) x n 2 + μ q [ n ] C T u 2 < 2 ,

Image (3.26)

0 < μ w [ n ] s n 2 + μ q [ n ] C T U i , n w n 2 < 2 ,

Image (3.27)

0 < μ q ( 1 ) [ n ] φ i 2 ( u n ( 2 ) ) Δ x ( 2 ) C T U i , n ( 1 ) w n 2 + μ q ( 2 ) [ n ] C T u n ( 2 ) 2 + μ w [ n ] φ i 2 ( u n ( 2 ) ) Δ x ( 2 ) s n 2 < 2 ,

Image (3.28)

0 < μ w ( 1 ) [ n ] φ i 2 ( u n ) Δ x X n w n ( 2 ) 2 + μ w ( 2 ) [ n ] r n 2 + μ q [ n ] C T U i , n w n ( 2 ) 2 < 2 .

Image (3.29)

The full derivation of these bounds can be found in [40,41,43].

3.4.2 Steady-State MSE

The aim of this paragraph is to evaluate the mean square performance of some of the proposed architectures at steady state. In particular, we are interested in the derivation of the EMSE of the nonlinear adaptive filter [2]. For simplicity we limit our analysis to the case of the Wiener and Hammerstein systems. The steady-state analysis is performed in a separate manner, by considering first the case of the linear filter alone and then the nonlinear spline function alone. Note that, since we are interested only in the steady state, the fixed subsystem in each substep can be considered adapted. The analysis is conducted by using a model of the same type of the used one, as shown in Fig. 3.5 for the particular case of the WSAF. Similar schemes are valid for the other kinds of architectures.

Image
Figure 3.5 Model used for statistical performance evaluation, in the particular case of the WSAF.

In the following we denote with ε[n]Image the a priori error of the whole system, with εw[n]Image the a priori error when only the linear filter is adapted while the spline control points are fixed and similarly with εq[n]Image the a priori error when only the control points are adapted while the linear filter is fixed. In order to make tractable the mathematical derivation, some assumption must be introduced:

A1. The noise sequence v[n]Image is independent and identically distributed with variance σv2Image and zero mean.

A2. The noise sequence v[n]Image is independent of xnImage, snImage, ε[n]Image, εw[n]Image and εq[n]Image.

A3. The input sequence x[n]Image is zero mean with variance σx2Image and at steady state it is independent of wnImage.

A4. At steady state, the error sequence e[n]Image and the a priori error sequence ε[n]Image are independent of φi(un)Image, φi2(un)Image, xnImage, snImage, CTunImage and CTUi,nwnImage.

The EMSE ζ is defined at steady state as

ζ lim n E { ε 2 [ n ] } .

Image (3.30)

Similar definitions hold for the EMSE ζwImage due to the linear filter component wnImage only and the EMSE ζqImage due to the spline nonlinearity qnImage only.

The MSE analysis has been performed by the energy conservation method [2]. Following the derivation presented in [41] and [44] we obtain an interesting property.

Proposition 1

The EMSE ζ of a Wiener and a Hammerstein spline adaptive filter satisfies the condition

ζ ζ w + ζ q .

Image (3.31)

The values of ζwImage and ζqImage are dependent on the particular SAF architecture considered. Specifically, for the Wiener case, we have [44]

ζ w = μ w [ n ] σ v 2 E { φ i 2 ( u n ) x n 2 } 2 Δ x 2 E { φ i ( u n ) c 3 q i , n } μ w [ n ] E { φ i 2 ( u n ) x n 2 } ,

Image (3.32)

where c3Image is the third row of the C matrix and

ζ q = μ q [ n ] σ v 2 E { C T u n 2 } 2 μ q [ n ] E { C T u n 2 } .

Image (3.33)

Other useful properties on the steady-state performances of the WSAF can be found in [44]. For the Hammerstein case, we have [41]

ζ w = μ w [ n ] σ v 2 E { s n 2 } 2 μ w [ n ] E { s n 2 }

Image (3.34)

and

ζ q = μ q [ n ] σ v 2 E { C T U i , n w n 2 } 2 μ q [ n ] E { C T U i , n w n 2 } .

Image (3.35)

3.5 Experimental Results

In order to validate the proposed nonlinear adaptive filtering solutions, several experiments were performed. In our previous works [4044], several convergence tests showed the effectiveness of the proposed architectures in converging to the optimal solution. Moreover, these works also reported some experimental tests in system identification scenarios.2

In this section, we provide two kinds of experimental results on the identification of nonlinear systems: (i) tests taken from the literature on the field and (ii) tests on data measured from real-world applications.

3.5.1 Results From Simulated Scenarios

Regarding the experiments on nonlinear systems taken from the literature, we compare results obtained by SAFs with those of a full third-order Volterra adaptive filter [59] and the approach proposed by Jeraj and Mathews in [60] where the nonlinearity is implemented as a third-order polynomial. In addition, Sandwich architectures are also compared with the LNL approach proposed by Hegde et al. in [61,62]. In all experiments the input signal x[n]Image consists of 50 000 samples of the signal generated by the following relationship:

x [ n ] = a x [ n 1 ] + 1 a 2 ξ [ n ] ,

Image (3.36)

where ξ[n]Image is a zero mean white Gaussian noise with unitary variance and a (0a<1Image) is a parameter that determines the level of correlation between adjacent samples. Experiments were conducted with a set to 0.1 and 0.95. In addition an additive white noise v[n]Image with variance σv2Image is considered.

In the following we propose four tests in order to verify the Wiener, Hammerstein, Sandwich 1 and Sandwich 2 architectures described in Section 3.3. All the SAF models use a CR-spline basis in (3.7) with Δx=0.2Image.

WSAF—The Wiener architecture to be identified is composed of the Back and Tsoi NARMA model reported in [63]. This model consists of a cascade of the third-order IIR filter

H ( z ) = 0.0154 + 0.0462 z 1 + 0.0462 z 2 + 0.0154 z 3 1 1.99 z 1 + 1.572 z 2 0.4583 z 3

Image (3.37)

and the following nonlinearity:

y [ n ] = sin ( x [ n ] ) .

Image (3.38)

The input signal x[n]Image is the colored signal obtained with (3.36), choosing a=0.95Image. The learning rates are set to μw=μq=0.1Image and the filter length is M=15Image taps. The third-order Volterra filter has Mv=10Image filter taps and the learning rate is set to μv=0.1Image, while the Jeraj and Mathews architecture is characterized by Ma=1Image, Mb=15Image, a nonlinearity of order N=3Image and learning rate μL=μNL=105Image. The MSE, averaged over 100 trials, is reported in Fig. 3.6A.

Image
Figure 3.6 Results of the proposed system identification tests for WSAF and HSAF. (A) MSE comparisons in the Wiener case. (B) MSE comparisons in the Hammerstein case.

HSAF—The Hammerstein architecture to be identified, drawn from [64], is composed of the following nonlinearity:

s [ n ] = 1 8 + 2 x [ n ] 1 ,

Image (3.39)

where Image is the floor operator, and the FIR filter h=12[1,0.75,0.5,0.25,0,0.25]Image. The learning rates are set to μw=0.01Image and μq=0.02Image, while the parameters of Volterra and Jeraj and Mathews architectures are the same as in the Wiener case. The MSE, averaged over 100 trials, is reported in Fig. 3.6B.

S1SAF—The identification of a Sandwich 1 system is drawn from [65]. The first and last blocks are the following two static nonlinear functions:

s [ n ] = x [ n ] 0.1 + 0.9 x 2 [ n ]

Image (3.40)

and

y [ n ] = r [ n ] 0.5 r 2 [ n ] + 0.2 r 3 [ n ] ,

Image (3.41)

while the second block is the following FIR filter:

H ( z ) = 1 + 0.5 z 0.25 z 1 + 0.05 z 2 + 0.001 z 3 .

Image (3.42)

The input signal x[n]Image is taken from (3.36) with a=0.1Image. Also SNR=30Image dB is considered. The learning rates are set to μw=μq(1)=μq(2)=0.05Image and the length is set to M=15Image. The proposed methods were compared with a standard full third-order Volterra filter, the approach of Jeraj and Mathews with the same parameters of the Wiener case and the LNL approach proposed by Hegde et al. with M1=M2=5Image, μA=μB=104Image and μC=103Image. The MSE, averaged over 100 trials, is reported in Fig. 3.7A.

Image
Figure 3.7 Results of the proposed system identification tests for S1SAF and S2SAF. (A) MSE comparisons in the Sandwich 1 case. (B) MSE comparisons in the Sandwich 2 case.

S2SAF—Finally, a Sandwich 2 to be identified is drawn from [66]. The first and last blocks are two fourth-order IIR filters, Butterworth and Chebyshev, respectively, with transfer functions

H B ( z ) = ( 0.2851 + 0.5704 z 1 + 0.2851 z 2 ) ( 1 0.1024 z 1 + 0.4475 z 2 ) ( 0.2851 + 0.5701 z 1 + 0.2851 z 2 ) ( 1 0.0736 z 1 + 0.0408 z 2 )

Image (3.43)

and

H C ( z ) = ( 0 . 2025 + 0.2880 z 1 + 0.2025 z 2 ) ( 1 1.01 z 1 + 0.5861 z 2 ) ( 0 . 2025 + 0.0034 z 1 + 0 . 2025 z 2 ) ( 1 0.6591 z 1 + 0.1498 z 2 ) ,

Image (3.44)

while the second block is the following nonlinearity:

y [ n ] = 2 x [ n ] 1 + | x [ n ] | 2 .

Image (3.45)

This system is similar to radio frequency amplifiers used in satellite communications (high-power amplifiers), in which the linear filters model the dispersive transmission paths, while the nonlinearity models the amplifier saturation. The input signal x[n]Image is as in (3.36) with a=0.1Image. The learning rates are set to μw(1)=0.005Image, μw(2)=0.009Image and μq=0.04Image and the lengths are set to M1=M2=5Image. Comparisons with Volterra, Jeraj and Mathews and Hegde et al. have been performed with the previous parameters. The MSE, averaged over 100 trials, is reported in Fig. 3.7B.

3.5.2 Results From Real-World Scenarios

In evaluating nonlinear architectures, it is important to test the performance of the proposed algorithm on completely unknown systems. For this purpose, we test our architecture from data sets collected from laboratory processes that can be described by block-oriented dynamic models. In particular, a number of interesting databases for system identification is represented by DaISy.3

From the DaISy database we have collected the following interesting and challenge data sets:

  • •  Exchanger [97-002]: the process is a liquid-saturated steam heat exchanger, where water is heated by pressurized saturated steam through a copper tube. The output variable is the outlet liquid temperature while the input variable is the liquid flow rate.
  • •  CD player arm [96-007]: data measured from the mechanical construction of a CD player arm. The input is the force of the mechanical actuators while the output is related to the tracking accuracy of the arm.
  • •  Thermic wall [96-011]: Heat flow density through a two-layer wall (brick and insulation layer). The input is the external temperature of the wall while the output is the heat flow density through the wall.
  • •  Heating system [99-001]: The experiment is a simple SISO heating system; the input drives a 300-Watt Halogen lamp, suspended several inches above a thin steel plate while the output is a thermocouple measurement taken from the back of the plate.
  • •  Continuous stirred tank reactor (CSTR) [98-002]: The process is a model of a CSTR, where the reaction is exothermic and the concentration is controlled by regulating the coolant flow. The input is the coolant flow while the output is the temperature.

All results have been compared with other state-of-the-art approaches used in previous experiments, namely a full third-order Volterra architecture [59] (with Mv=15Image), the LNL approach proposed by Hegde et al. in [61,62] (with M1=M2=5Image), where the nonlinearity is implemented as a third-order polynomial and the approach proposed by Jeraj and Mathews in [60] (with Ma=1Image, Mb=5Image and N=3Image). In order to provide a fair comparison of the performance of such algorithms, Table 3.2 shows their asymptotic computational complexity.

Table 3.2

Comparison of the asymptotic computational complexity of the algorithms used in experimental comparisons
Algorithm Asymptotic complexity
WSAF [40] O ( M ) Image
HSAF [41] O ( M ) Image
S1 SAF [43] O ( M ) Image
S2 SAF [43] O ( M 1 M 2 ) Image
Third-order Volterra [59] O ( M v 3 ) Image
Hegde et al.[61] O ( M 1 M 2 ) Image
Jeraj and Mathews [60] O ( ( M b M a N ) 2 ) Image

Moreover, we have chosen two additional measured data from coupled electric drives that consist of two electric motors that drive a pulley using a flexible belt and two measured data from a fluid level control system consisting of two cascaded tanks with free outlets fed by a pump. Additional details as regards these systems and data sets can be found in [67].

A last data set is taken from [68,69] and consist of the identification of a nonlinear electric circuit composed of the cascade of a third-order Chebyshev linear filter, a nonlinearity built by using a diode circuit and a second linear filter designed as a third-order inverse Chebyshev filter with a transmission of zero in the frequency band of interest.

Results in terms of MSEs for all data sets are summarized in Table 3.3. It can be seen that the proposed SAF approaches provide the best performance in almost all cases.

Table 3.3

MSE in dB of the SAF architectures, compared with other referenced approaches, using real-world data

WSAF HSAF S1SAF S2SAF Volterra Jeraj & Mathews Hedge et al.
Exchanger −34.18 −25.12 −39.14 −32.38 −16.66 −7.77 −7.87
CD player arm −13.89 −10.68 −19.94 −12.50 −19.23 −5.53 −5.58
Thermic wall −16.29 −15.66 −22.71 −16.27 −24.26 −7.58 −7.66
Heating system −17.99 −17.94 −22.09 −17.87 −8.79 −16.04 −16.02
CSTR −39.53 −42.03 −53.85 −41.42 −21.51 −26.16 −26.29
Drive 1 −15.81 −22.52 −12.51 −13.29 −13.51 −2.33 −2.55
Drive 2 −14.09 −2.65 −19.17 −15.25 −6.54 −9.59 −8.34
Tank 1 −1.08 −17.33 −1.13 −21.60 −10.81 −14.62 −18.54
Tank 2 −2.65 −14.59 −5.70 −20.49 −2.50 −11.84 −15.39
Diode −15.88 −11.69 −11.98 −19.79 −7.02 −14.02 −17.25

Image

3.5.3 Applicative Scenarios

In the previous subsections, we provided some numerical results on the general problem of nonlinear system identification. However, SAFs have been used in applicative scenarios providing interesting and very good results.

A first applicative scenario is the nonlinear acoustic echo cancellation (NAEC) problem. The SAF approach has been successfully applied to NAEC in [30], where it is shown that the proposed solutions outperform other linear and nonlinear approaches.

Another challenging scenario is the nonlinear adaptive noise cancellation (NANC) problem. In particular, in [45] authors show the effectiveness of SAF approach to other state-of-the-art algorithms. Moreover, the proposed solution has been extended with additional refinements in [46] and [47].

Finally, as underlined in Section 3.3.5, the SAF approach has been recently extended to the diffusion case [50], involving a distributed (wireless) sensors network. The authors of this work derive a D-SAF by applying ideas from the DA framework, as explained in Chapter 9. Experimental evaluations on the D-SAF approach show that it is able to robustly learn the underlying nonlinear model, with a significant gain compared with a noncooperative solution.

3.6 Conclusion

In this chapter, a complete overview of a recent class of adaptive nonlinear filters, SAFs, has been provided. Specifically, we have described the basic block-oriented architecture implemented by SAFs, namely the Wiener, Hammerstein, NLN and LNL ones, as well as the related learning rules. Additionally, we have referred to the main theoretical results found in the literature on SAFs; the bounds on the choice of the learning rates and the theoretical evaluation of the EMSE for the Wiener and Hammerstein architectures.

Finally, after the introduction of some possible real-world applicative scenarios, some numerical results have shown the effectiveness of the proposed architectures.

References

[1] A. Uncini, Fundamentals of Adaptive Signal Processing, Signals and Communication Technology. Switzerland: Springer International Publishing; 2015 10.1007/978-3-319-02807-1.

[2] A.H. Sayed, Fundamentals of Adaptive Filtering. John Wiley & Sons; 2003.

[3] S. Haykin, Adaptive Filter Theory. 4th edition Prentice-Hall; 2002.

[4] L. Garcia, ed. Adaptive Filtering Applications. InTech; 2011.

[5] N. Wiener, Nonlinear Problems in Random Theory. John Wiley & Sons; 1958.

[6] O. Montiel, O. Castillo, P. Melin, R. Sepúlveda, Black box evolutionary mathematical modeling applied to linear systems, International Journal of Intelligent Systems 2005;20(2):293–311.

[7] R.K. Pearson, M. Pottmann, Gray-box identification of block-oriented nonlinear models, Journal of Process Control 2000;10(4):301–3115.

[8] V.J. Mathews, G.L. Sicuranza, Polynomial Signal Processing. John Wiley & Sons; 2000.

[9] D. Mirri, G. Iuculano, F. Filicori, G. Pasini, G. Vannini, G. Pellegrini Gualtieri, A modified Volterra series approach for nonlinear dynamic systems modeling, IEEE Transactions on Circuits and Systems—I: Fundamental Theory and Applications 2002;49(8):1118–1128.

[10] S. Haykin, Neural Networks and Learning Machines. 2nd edition Pearson Publishing; 2009.

[11] J. de Jesus Rubio, W. Yu, Stability analysis of nonlinear system identification via delayed neural networks, IEEE Transactions on Circuits and Systems—II: Express Briefs 2007;54(2):161–165.

[12] L. Ljung, System Identification—Theory for the User. 2nd edition Upper Saddle River, N.J.: Prentice Hall; 1999.

[13] A.E. Nordsjo, L.H. Zetterberg, Identification of certain time-varying Wiener and Hammerstein systems, IEEE Transactions on Signal Processing 2001;49(3):577–592.

[14] S. Boyd, L.O. Chua, Uniqueness of a basic nonlinear structure, IEEE Transactions on Circuits and Systems 1983;CAS-30(9):648–651.

[15] R. Haber, H. Unbehauen, Structure identification of nonlinear dynamic systems—a survey on input/output approaches, Automatica 1990;26(4):651–677.

[16] E.W. Bai, Frequency domain identification of Wiener models, Automatica 2003;39:1521–1530.

[17] M. Korenberg, I. Hunter, Two methods for identifying Wiener cascades having noninvertible static nonlinearities, Annals of Biomedical Engineering 1999;27(6):793–804.

[18] D. Westwick, M. Verhaegen, Identifying MIMO Wiener systems using subspace model identification methods, Signal Processing 1996;52(2):235–258.

[19] G. Mzyk, A censored sample mean approach to nonparametric identification of nonlinearities in Wiener systems, IEEE Transactions on Circuits and Systems—II: Express Briefs 2007;54(10):897–901.

[20] E.W. Bai, Frequency domain identification of Hammerstein models, IEEE Transactions on Automatic Control 2003;48:530–542.

[21] I.J. Umoh, T. Ogunfunmi, An affine projection-based algorithm for identification of nonlinear Hammerstein systems, Signal Processing 2010;90(6):2020–2030.

[22] J. Vörös, Identification of Hammerstein systems with time-varying piecewise-linear characteristics, IEEE Transactions on Circuits and Systems—II: Express Briefs 2005;52(12):865–869.

[23] L. Ljung, Identification of nonlinear systems, 9th International Conference on Control, Automation, Robotics and Vision. ICARCV 2006, Singapore. 2006:1–9.

[24] H.-E. Liao, W.A. Sethares, Cross-term analysis of LNL models, IEEE Transactions on Circuits and Systems—I: Fundamentals Theory and Applications 1996;43(4):333–335.

[25] S. Billings, S.Y. Fakhouri, Identification of a class of nonlinear systems using correlation analysis, Proceedings of IEEE 1978;125(7):691–697 10.1049/piee.1978.0161.

[26] O. De Feo, M. Storace, PWL identification of dynamical systems: some examples, Proceedings of the 2004 International Symposium on Circuits and Systems, ISCAS2004, vol. 4. 2004:665–668.

[27] C. Wen, X. Ma, A basis-function canonical piecewise-linear approximation, IEEE Transactions on Circuits and Systems—I: Regular Papers 2008;55(5):1328–1334 10.1109/TCSI.2008.916552.

[28] B.N. Bond, Z. Mahmood, Y. Li, R. Sredojević, A. Megretski, V. Stojanović, Y. Avniel, L. Daniel, Compact modeling of nonlinear analog circuits using system identification via semidefinite programming and incremental stability certification, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems 2010;29(8):1149–1162 10.1109/TCAD.2010.2049155.

[29] S. Orcioni, Approximation and analog implementation of nonlinear systems represented by Volterra series, 6th IEEE International Conference on Electronics, Circuits and Systems. 1999:1109–1113.

[30] M. Scarpiniti, D. Comminiello, R. Parisi, A. Uncini, Comparison of Hammerstein and Wiener systems for nonlinear acoustic echo cancelers in reverberant environments, Proc. of 17th International Conference on Digital Signal Processing. DSP2011, Corfu, Greece. 2011:1–6 10.1109/ICDSP.2011.6004959.

[31] K.J. Hunt, M. Munih, N.N. Donaldson, F.M.D. Barr, Investigation of the Hammerstein hypothesis in the modeling of electrically stimulated muscle, IEEE Transactions on Biomedical Engineering 1998;45(8):99–1009.

[32] Z. Zhu, H. Leung, Adaptive identification of nonlinear systems with application to chaotic communications, IEEE Transactions on Circuits and Systems—I: Fundamental Theory and Applications 2000;47(7):1072–1080.

[33] L. Peng, H. Ma, Design and implementation of software-defined radio receiver based on blind nonlinear system identification and compensation, IEEE Transactions on Circuits and Systems—I: Regular Papers 2011;58(11):2776–2789 10.1109/TCSI.2011.2151050.

[34] B. Kwak, A.E. Yagle, J.A. Levitt, Nonlinear system identification of hydraulic actuator friction dynamics using a Hammerstein model, Proceedings of the IEEE ASSP'98, vol. 4. Seattle, WA. 1998:1933–1936.

[35] J.-C. Jeng, M.-W. Lee, H.-P. Huang, Identification of block-oriented nonlinear processes using designed relay feedback tests, Industrial & Engineering Chemistry Research 2005;44(7):2145–2155.

[36] O. De Feo, M. Storace, Piecewise-linear identification of nonlinear dynamical systems in view of their circuit implementations, IEEE Transactions on Circuits and Systems—I: Regular Papers 2007;54(7):1542–1554 10.1109/TCSI.2007.899613.

[37] C. Wen, H. Ma, A canonical piecewise-linear representation theorem: geometrical structures determine representation capability, IEEE Transactions on Circuits and Systems—II: Express Briefs 2011;58(12):936–940 10.1109/TCSII.2011.2172715.

[38] A. Stenger, W. Kellerman, Adaptation of a memoryless preprocessor for nonlinear acoustic echo cancelling, Signal Processing 2000;80(9):1747–1760.

[39] C. de Boor, A Practical Guide to Splines. revised edition Springer-Verlag; 2001.

[40] M. Scarpiniti, D. Comminiello, R. Parisi, A. Uncini, Nonlinear spline adaptive filtering, Signal Processing 2013;93(4):772–783 10.1016/j.sigpro.2012.09.021.

[41] M. Scarpiniti, D. Comminiello, R. Parisi, A. Uncini, Hammerstein uniform cubic spline adaptive filters: learning and convergence properties, Signal Processing 2014;100:112–123 10.1016/j.sigpro.2014.01.019.

[42] M. Scarpiniti, D. Comminiello, R. Parisi, A. Uncini, Nonlinear system identification using IIR spline adaptive filters, Signal Processing 2015;108:30–35 10.1016/j.sigpro.2014.08.045.

[43] M. Scarpiniti, D. Comminiello, R. Parisi, A. Uncini, Novel cascade spline architectures for the identification of nonlinear systems, IEEE Transactions on Circuits and Systems—I: Regular Papers 2015;62(7):1825–1835 10.1109/TCSI.2015.2423791.

[44] M. Scarpiniti, D. Comminiello, G. Scarano, R. Parisi, A. Uncini, Steady-state performance of spline adaptive filters, IEEE Transactions on Signal Processing 2016;64(4):816–828 10.1109/TSP.2015.2493986.

[45] V. Patel, N.V. George, Nonlinear active noise control using spline adaptive filters, Applied Accoustics 2015;93:38–43 10.1016/j.apacoust.2015.01.009.

[46] V. Patel, D. Comminiello, M. Scarpiniti, N.V. George, A. Uncini, Design of hybrid nonlinear spline adaptive filters for active noise control, Proc. of 2016 International Joint Conference on Neural Networks. IJCNN2016, Vancouver, Canada. 2016:3420–3425 10.1109/IJCNN.2016.7727637.

[47] V. Patel, N.V. George, Compensating acoustic feedback in feed-forward active noise control systems using spline adaptive filters, Signal Processing 2016;120:448–455 10.1016/j.sigpro.2015.10.003.

[48] S. Guan, Z. Li, Normalised spline adaptive filtering algorithm for nonlinear system identification, Neural Processing Letters 2017:1–13 10.1007/s11063-017-9606-6.

[49] S. Peng, Z. Wu, X. Zhang, B. Chen, Nonlinear spline adaptive filtering under maximum correntropy criterion, 2015 IEEE Region 10 Conference. TENCON 2015. 2015:1–5 10.1109/TENCON.2015.7373051.

[50] S. Scardapane, M. Scarpiniti, D. Comminiello, A. Uncini, Diffusion spline adaptive filtering, Proc. of the 24-th European Signal Processing Conference. EUSIPCO2016, Budapest, Hungary. 2016:1498–1502 10.1109/EUSIPCO.2016.7760498.

[51] L. Schumaker, Spline Functions: Basic Theory. 3rd edition Krieger Publishing Company; 1993.

[52] H. Prautzsch, W. Boehm, M. Paluszny, Bézier and B-Spline Techniques. 1st edition Springer-Verlag; 2002.

[53] S. Guarnieri, F. Piazza, A. Uncini, Multilayer feedforward networks with adaptive spline activation function, IEEE Transaction on Neural Network 1999;10(3):672–683.

[54] R.H. Bartels, J.C. Beatty, B.A. Barsky, An Introduction to Splines for Use in Computer Graphics & Geometric Modeling. Morgan Kaufmann; 1987.

[55] E. Catmull, R. Rom, A class of local interpolating splines, R.E. Barnhill, R.F. Riesenfeld, eds. Computer-Aided Geometric Design. New York Academic; 1974:317–326.

[56] D. Comminiello, M. Scarpiniti, L.A. Azpicueta-Ruiz, J. Arenas-Garcìa, A. Uncini, Combined nonlinear filtering architectures involving sparse functional link adaptive filters, Signal Processing 2017;135:168–178 10.1016/j.sigpro.2017.01.009.

[57] M. Rathod, V. Patel, N.V. George, Generalized spline nonlinear adaptive filters, Expert Systems with Applications 2017;48:122–130 10.1016/j.eswa.2017.04.043.

[58] P. Borwein, T. Erdelyi, Polynomials and Polynomial Inequalities. Springer-Verlag; 1995.

[59] M. Schetzen, Nonlinear system modeling based on the Wiener theory, Proceedings of the IEEE 1981;69(12):1557–1573.

[60] J. Jeraj, V.J. Mathews, A stable adaptive Hammerstein filter employing partial orthogonalization of the input signals, IEEE Transactions on Signal Processing 2006;54(4):1412–1420.

[61] V. Hegde, C. Radhakrishnan, D.J. Krusienski, W.K. Jenkins, Architectures and algorithms for nonlinear adaptive filters, Conference Record of the 36th Asilomar Conference on Signals, Systems and Computers, vol. 2. Pacific Grove, CA, USA. 2002:1015–1018 10.1109/ACSSC.2002.1196937.

[62] V. Hegde, C. Radhakrishnan, D.J. Krusienski, W.K. Jenkins, Series-cascade nonlinear adaptive filters, Proceedings of the 45th Midwest Symposium on Circuits and Systems, MWSCAS2002, vol. 3. Tulsa, Oklahoma. 2002:219–222 10.1109/MWSCAS.2002.1187010.

[63] A.D. Back, A.C. Tsoi, A simplified gradient algorithm for IIR synapse multilayer perceptrons, Neural Computation 1993;5(3):456–462.

[64] Z. Hasiewicz, M. Pawlak, P. Sliwinski, Nonparametric identification of nonlinearities in block-oriented systems by orthogonal wavelets with compact support, IEEE Transactions on Circuits and Systems—I: Regular Papers 2005;52(2):427–442.

[65] Y. Zhu, Estimation of an N–L–N Hammerstein–Wiener model, Automatica 2002;38(9):1607–1614.

[66] T.M. Panicker, V.J. Mathews, G.L. Sicuranza, Adaptive parallel-cascade truncated Volterra filters, IEEE Transactions on Signal Processing 1998;46(10):2664–2673.

[67] T. Wigren, J. Schoukens, Three free data sets for development and benchmarking in nonlinear system identification, Proc. of 2013 European Control Conference. ECC2013, Zürich, Switzerland. 2013:2933–2938.

[68] Y. Han, R.A. de Callafon, Identification of Wiener–Hammerstein benchmark model via rank minimization, Control Engineering Practice 2012;20(11):1149–1155 10.1016/j.conengprac.2012.05.004.

[69] J. Schoukens, J. Suykens, L. Ljung, Wiener–Hammerstein benchmark, 15th IFAC Symposium on System Identification. SYSID 2009, St. Malo, France. 2009:1–4.


1  “In this work we consider only real-valued variables.”

2  “Some Matlab source codes, implementing the SAF architectures and experimental results, are available at the following web page: https://github.com/ispamm/SAF-Toolbox.”

3  “Available at http://homes.esat.kuleuven.be/~smc/daisy/daisydata.html.”

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
18.222.196.175