Theorem 1.2 If each point between the noise and between different dimensions, and between (Wi,Wi)'(Wi,Wi) and the noise are independent of each other, the noise has a mean value with the same variance [14]. Then, δ(W,W)=(Wi,Wi)'(Wi,Wi)δ(W,W)=(Wi,Wi)'(Wi,Wi) can be estimated as

where ‖⋅‖ is the Euclidean norm; (δ,δ)2=Dj=1(δi,δi)2,(δi,δi)=Var(εi,εi)(δ,δ)2=j=1D(δi,δi)2,(δi,δi)=Var(εi,εi)(i=1,2,...,D),(εi,εi)(i=1,2,...,D),(εi,εi) represents the i th component of (ε,ε),l=rank(U(xi)),(ε,ε),l=rank(U(xi)),U(xi)),and(λ,λ)minU(xi)),and(λ,λ)min is the smallest nonzero eigenvalue of (U(xi),U(xi))T(U(xi),U(xi))T(U(xi),U(xi)).(U(xi),U(xi)).

From this theorem, it can be concluded that when the neighbourhood size k is known, the error of (, ) is mainly determined by three factors: (1) the influence of noise interference, that is, the size of (δ,δ);(δ,δ); (2) neighbourhood effects, that is, the size of (λ,λ)min(λ,λ)min and the rank l; and (3) the impact of weight energy, that is, the size of (W,W)'.(W,W)'.

1.2.2.3Improved locally linear embedding (ILLE)

Determining the size of the neighbourhood k is a very important problem. If k is too small, the mapping will cause the sub-manifolds to break away from the continuous manifold and not reflect any global properties. If k is too large, the entire dataset will be considered as the local neighbourhood, and the map will lose its nonlinear characteristic [15]. In this section, we propose an ILLE method that uses the diffusion and growing self-organizing map (DGSOM) [16] to define neighbourhood relations. For example, if a point has five associated points in the DGSOM map, the five data points are considered neighbours of that point. Thus, the size of the neighbourhood of each sample point, k, is different, which overcomes the shortcoming of the traditional LLE method that the neighbourhood size of each point is fixed. In addition, the ILLE reduces the influence of noise by reducing the noise interference and weighting energy. By way of example, we show that ILLE can adapt to the neighbourhood and overcome the sensitivity problem, enabling it to solve the error problem caused by the three factors identified above.

The core of the ILLE method is to seek the k-dimensional dynamic fuzzy vector (Wi,Wi)=((W1i,W1i),(W2i,W2i),...,(Wki,Wki)).(Wi,Wi)=((W1i,W1i),(W2i,W2i),...,(Wki,Wki)). The optimization problem is

which is to find the corresponding dynamic fuzzy topology matrix (, ), and finally get (MM, MM).

ILLE avoids the ill-posedness of LLE by introducing the dynamic fuzzy regularization term (g,g)(kl=1(wli,wli)),(g,g)(l=1k(wli,wli)), which reduces the influence of noise interference. This term must satisfy two points: (1) (1.5) must have a steady solution; and (2) kl=1(wli,wli)l=1k(wli,wli) in (1.5) must be minimized, typically such that (g,g)(x,x)=(g,g)(x,x)=(x,x)2,(η,η)(x,x)2,(η,η) is the improvement factor, which is related to the neighbourhood number k, the dimension of the dynamic fuzzy dataset D, and the dimension of the dynamic fuzzy dataset. We can see from the calculation that the choice of (η,η)(η,η) does not have a large effect on the outcome. The description of ILLE is shown in Algorithm 1.1.

Algorithm 1.1 ILLE algorithm

Input: (x,x)={(x1,x1),(x2,x2),...,(xn,xn)}(RD×n,RD×n),(x,x)={(x1,x1),(x2,x2),...,(xn,xn)}(RD×n,RD×n),(xj,xi)(RD,RD)(i=1,2,...,n)(xj,xi)(RD,RD)(i=1,2,...,n)

Output:(Y,Y)={(Y1,Y1),(Y2,Y2),(Yn,Yn)}(Rd×n,Rd×n),(Y,Y)={(Y1,Y1),(Y2,Y2),(Yn,Yn)}(Rd×n,Rd×n),(yj,yj)(Rd,Rd)(l=1,2,...,n)(yj,yj)(Rd,Rd)(l=1,2,...,n)

For each (xj,xj)(xj,xj) do

1.Choose neighbourhood: The k neighbourhoods {(x1i,x1i)|=1,2,...,k}{(x1i,x1i)=1,2,...,k} of each point (xi,xi)(xi,xi) are determined according to the DGSOM method, where k may be different for each neighbourhood. A linear combination of {(x1j,x1j)|l=1,2,...,k}{(x1j,x1j)l=1,2,...,k} is used to denote (xi,xi),(xi,xi), that is, the hyperboloid passing through {(x1j,x1j)|l=1,2,...,k}{(x1j,x1j)l=1,2,...,k} is used to approximate the local block near {(xij,xij)|l=1,2,...,k}{(xij,xij)l=1,2,...,k} in the manifold.

2.Get weight: For each (xi,xi),find(Wi,Wi)=((W1j,W1j),(W2j,W2j),...,(Wkj,Wkj))(xi,xi),find(Wi,Wi)=((W1j,W1j),(W2j,W2j),...,(Wkj,Wkj)) that minimizes (Q,Q)(wi,wi)=(xi,xi)ki=1(wli,wli)(x1i,x1i)2+(η,η)kl=1(w1i,w1i)2.(Q,Q)(wi,wi)=(xi,xi)i=1k(wli,wli)(x1i,x1i)2+(η,η)l=1k(w1i,w1i)2. and finally obtain the dynamic fuzzy topology matrix (, ).

3. Reduce dimension: Minimize the loss function

where (Y,Y)=((Y1,Y2),(Y2,Y2),...,(Yn,Yn))(Rd×n,Rd×n),(yi,yi)(Rd,Rd)(i=)(Y,Y)=((Y1,Y2),(Y2,Y2),...,(Yn,Yn))(Rd×n,Rd×n),(yi,yi)(Rd,Rd)(i=) 1,2,. . .n). To make the solution unique, we add the constraint to (Yi,Yi)(Rd,Rd)(i=(Yi,Yi)(Rd,Rd)(i= 1,2,. . .n):

where (I,I)d×d(I,I)d×d is the dynamic fuzzy unit matrix.

The computational complexity of each step of ILLE is as follows. The number of computations in steps 1–3 is O(Dn2, ODnk3, and Odn2, respectively. The computational complexity of ILLE is similar to that of LLE, but ILLE is more robust against noise interference.

1.2.2.4DFML algorithm and its stability analysis

A. DFML Algorithm [5, 7, 8, 11]

Consider the DFMLS set up by the following dynamic fuzzy rules:

Rl: IF (x,x)(k)is(Al1,Al1)(x,x)(k)is(Al1,Al1) is (x,x)(k1)(x,x)(k1) and (Al2,Al2)(Al2,Al2) is (x,x)(x,x) and ...and (x,x)(x,x)(kn+1)(kn+1) is (A1n,A1n)(A1n,A1n) and (u,u)(k)is(Bl,Bl)(u,u)(k)is(Bl,Bl) THEN

where (X,X)(k)=((x,x)(k),(x,x)(k1),...,(x,x)(kn+1))T(U,U)(X,X)(k)=((x,x)(k),(x,x)(k1),...,(x,x)(kn+1))T(U,U) is the system state vector, (Al,Al)=((al1,al1),(al2,al2),...,(aln,aln)),(ali,ali)(Al,Al)=((al1,al1),(al2,al2),...,(aln,aln)),(ali,ali) (i = 1, 2, ..., n) is the corresponding dynamic fuzzy constant, and (Ali,Ali)(Ali,Ali)(R,R)n×n,(Bt,Bt)(R,R)n×n(R,R)n×n,(Bt,Bt)(R,R)n×n is the dynamic fuzzy constant matrix. The linear state equation corresponding to each dynamic fuzzy rule is called the DFML subsystem.

For the problem of noise in the learning process, we use the robust factor proposed in [17] to reduce the interference:

where (ei,ej)=(yd,yd)(yi,yi),(yd,yd)(ei,ej)=(yd,yd)(yi,yi),(yd,yd) is the expectation output, (yi,yi)(yi,yi) is the actual output, and (ci,ci)(ci,ci) is the central value of the confidence interval for the statistical estimation of the deviation (Stdak,cur,Stdak,cur)(Stdak,cur,Stdak,cur)(Gak,cur,Gak,cur)(Gak,cur,Gak,cur) is a nonnegative integrable function of (Ck,Ck):(Ck,Ck): and

The learning algorithm of the DFMLS is shown in Algorithm 1.2.

Algorithm 1.2 Dynamic fuzzy machine learning

Input: (x,x)(x,x)

Output: (yk,yk)(yk,yk)

//(uAjl,uAjl),(ubjl,ubjl)//(uAjl,uAjl),(ubjl,ubjl) is the corresponding dynamic fuzzy membership, (yd,yd)(yd,yd) is the desired output //

(1)Use the ILLE method to reduce the dimension of the DFD;

(2)

(3)(v,v)l=nj=1(uA1j,uA1j)[(x,x)(kl+1)](ub1j,ub1j)[(u,u)(k)];(v,v)l=j=1n(uA1j,uA1j)[(x,x)(kl+1)](ub1j,ub1j)[(u,u)(k)];

(4)(yk,yk)=kj1(hj,hj)(x,x)(i)kj1(hj,hj);(yk,yk)=j1k(hj,hj)(x,x)(i)j1k(hj,hj);

(5)(ek,ek)=(yd,yd)(yk,yk);(ek,ek)=(yd,yd)(yk,yk);

(6)(ϕ,ϕ)((ek,ek)(ck,ck))=2.exp((λk,λk)|(ek,ek)(ck,ck)|);(ϕ,ϕ)((ek,ek)(ck,ck))=2.exp((λk,λk)(ek,ek)(ck,ck));//(λk,λk)and(ck,ck)//(λk,λk)and(ck,ck) are determined as in [17] //

(7)(p,p)(k)=12kj=1|(ei,ei)|2;//performance//(p,p)(k)=12j=1k(ei,ei)2;//performance//

(8)if|(ek,ek)>(ε,ε)|.if(ek,ek)>(ε,ε).

//(ε,ε)(ε,ε) is the maximum tolerance error, its value can be based on system requirements or experience. If the learning error is greater than the maximum tolerance, then the learning is invalid//

then Discard (yk,yk),(yk,yk), feedback to the rule base, notify the adjustment rules of the parameters, perform a parametric learning algorithm.

(u,u)(k)=(u,u)(k)+α(ek,ek)(u,u)(k)=(u,u)(k)+α(ek,ek) //α is the gain learning coefficient

Continue from scratch

else if | (ek,ek)(δ,δ)|(ek,ek)(δ,δ)

//(δ,δ)(δ,δ) is an acceptable error, much smaller than (ε,ε),(ε,ε),, its value can be given according to system requirements or experience. If the error after learning is less than (or equal to) this acceptable level, it is considered to have reached the accuracy of the learning requirements //

then The learning process is finished, corrected, and output

(yk,yk)=(yk,yk)+β(ϕ,ϕ)((ek,ek)(ck,ck))=(yk,yk)//β(yk,yk)=(yk,yk)+β(ϕ,ϕ)((ek,ek)(ck,ck))=(yk,yk)//β is the correction factor

else Feedback to the rule base, notify the adjustment rules in the parameters, and implement the parameter learning algorithm.

Continue from scratch;

end if

end if

Note: (hj,hj)(hj,hj) is the corresponding weight coefficient, and the membership function is generally taken as the weight coefficient.

B. Stability analysis

Denoting (, )(k) as ((x,x)(k),(x,x)(k1),...,(x,x)(kn+1))T(U,U)((x,x)(k),(x,x)(k1),...,(x,x)(kn+1))T(U,U) and

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
18.223.171.162