The state feedback gain of subsystems 1 and 2 is then:

(K1,K1)=[(211.3,211.3)(45.3,45.3)](K2,K2)=[(5607.5,5607.5)(1527.9,1527.9)]

and so:

(A0,A0)=[(0,0)(1,1)(13.325,13.325)(0,0)](C0,C0)=[(0,0)(1,1)(26.27,26.27)(9.7,9.7)](M,M)(M,M)T=[(0.0001,0.0001)(0,0)(0,0)(4.83,4.83)](N,N)=[(0,0)(0,0)(1.952,1.952)(0,0)(0,0)(0.01,0.01)(0,0)(1.01,1.01)]T.

Finally, a dynamic fuzzy positive definite matrix is obtained:

(P,P)=[(0.013,0.013)(1.012,1.012)(27.462,27.462)(10.05,10.05)].

Therefore, from Theorem (1.28), the inverted pendulum system is asymptotically stable under the control of the dynamic fuzzy learning controller (u,u)(k)=((u1,u1)(k)(K1,K1)+(u2,u2)(k)(K2,K2))(x,x)(k).

Fig. 1.11: (x1,x1)(0)=(60°,60°),(x2,x2)(0)=(0,0) simulation curve.

The initial points

(x1,x1)(0)=(60,60),(x2,x2)(0)=(0,0)(x1,x1)(0)=(89,89),(x2,x2)(0)=(0,0)

give the simulation curves shown in Figs. 1.11 and 1.12, respectively. It can be seen from these figures that the system stabilizes at equilibrium point (0,0) after 1.01s; i.e. point (0,0) is the stable equilibrium of the closed-loop system. The inverted pendulum system in [37] becomes stable at (0,0) after 1.05s, so the performance of the proposed system is superior.

1.5Dynamic fuzzy relational learning algorithm

In machine learning systems, data with dynamic fuzzy relations are universal. Dynamic fuzzy relations provide an effective method for describing and dealing with dynamic fuzzy phenomena. For further information on dynamic fuzzy relations, see Appendix 8.2 and References [1, 2, 4]. In this section, a learning algorithm for dynamic fuzzy relation (̄R, ̄R) is described based on the theory of dynamic fuzzy relations.

Fig. 1.12. (x1,x1)(0)=(89°,89°),(x2,x2)=(0,0) simulation curve.

1.5.1An outline of relational learning

Relational learning is an important paradigm in the field of machine learning. Its early work can be traced back to the famous Winston program of 1970, which could learn concepts such as arches. This program uses the semantic web as a descriptive language. It takes the arches with their components and properties and considers the description of the nature of properties as a node of the semantic network. In the same year, Codd published a landmark paper on “A Relational Model of Data for Large Shared Data Banks”, which was a prelude to the idea of relational databases. This paper is cited in many applications of relational learning, as it enables data with complex internal relations to be concisely expressed.

As we know, early studies on relational learning focus on inductive logic programming (ILP) [38, 39]. Although the logical representations that are presented in Baner’s “A Language for the Description of Concepts” and in Michalski’s AQ algorithm are first applied to learning problems, the foundation of this work is the first-order predicate formula presented by Plotkin in the 1970s, which laid the theoretical foundation for the development of ILP. The thrust of Plotkin’s work is to use the “special to general” approach to obtain a conclusion; this is strictly limited in practice, as a generalization of the smallest general conclusion is not finite. Thus, Shapiro used the general to special method to research the induction problem on the Horn clause set and proposed the MIS learning system in 1983. His work clarified the relationship between the logical program and inductive learning.

Early research on ILP focuses mainly on relationship rule induction, i.e. first-order logic concept learning and logic program composition. Representative systems include CIGOL (Muggleton, 1988) [54], FOIL (Quinlan, 1990) [55], GOLEM (Muggleton, 1990) [56], and LINUS (Lavarac, Dzeroski, and Grobelnik, 1991) [57]. In recent years, the research scope of ILP has been expanded to almost all learning tasks, such as classification, regression, clustering, and correlation analysis, and has generated relational classification rules, relational regression trees, and relational learning based on distance and relationship-related rules. Representative systems include PROLOG (Muggleton, 1995) [58] and ALEPH (Srinivasan, 2000) [59]. Later, the characteristics of statistical learning theory and gradual research on Statistical Relational Learning (SRL) led to many relational learning methods based on SRL, such as SRL based on Bayesian networks, SRL based on Markov networks, SRL based on hidden Markov models, and SRL based on random grammar (Popescul, Ungar, Lawrence, & Pennock, 2003) [40, 41].

(1)SRL based on Bayesian networks

Ng and Subrahmanian proposed the first framework for Probabilistic Logic Programs (PLPs) [42] in 1992. This framework enabled the description of probabilistic information within the fixed-point semantics of PLP, which is a valuable tool for both probabilistic information and common sense reasoning. In 1997, Ngo and Haddawy combined the logic with Bayesian networks by defining probabilities based on first-order logic or relational interpretation. Their work enriched the PLP content. In 1996, Debabrata and Sumit Sarkar proposed the Probabilistic Relational Model (PRM) [43], which deals with uncertain information. They systematically described the theory of probabilistic databases and defined a series of concepts such as super keys, primary keys, candidate keys, probabilistic patterns, probabilistic relations, and probability relational databases. They also defined a whole train of operations such as projection, selection, difference, and renaming, and discussed how to realize probabilistic databases and the NULL problem. Their work also converted the partial probability distribution to the interval estimation and point estimation. PRM extended the standard Bayesian network using an entity relationship model as a basic representation framework. Today, PRM has successfully solved many problems in relational clustering, hypertext classification, and so on. In 2000, Kersting and De Raedt developed Bayesian Logic Programs (BLPs) [44] by combining logic programs with Bayesian networks. BLPs not only handle Boolean variables but also take continuous-valued variables. They can solve a number of instances of clauses that have the same head using a simplified form of the combination of rules.

(2)SRL based on Markov networks

In 2002, Taskar, Abbeel, and Koller proposed the Relational Markov Network (RMN) [45] by introducing the relationship between entities into the Markov network. RMN extended the Markov network to relational databases and has been used to classify relational data. RMN defined the groups and potential functions between attributes of related entities at the template level, thus giving an arbitrary set of examples a distribution. In 2004, Richardson and Domingos proposed a Markov logic network (MLN) [46], which is a closed Markov network generated by a first-order logic knowledge database with a weighted value as a template. The advantage of MLN is that it introduced domain knowledge into Markov networks while providing a simple description language for large-scale networks. Additionally, MLN increased the uncertainty processing capability for first-order logic and served as a unified framework for many SRL tasks.

(3)RSL based on hidden Markov models

In 2002, Anderson and Domingos proposed a Relational Markov Model (RMM) [47] by introducing relations (logic) into Markov models. RMM allows state variable values to be of multiple types and uses one predicate or relationship to represent the same type of states. Thus, it overcomes the shortcoming of state variables whereby the Markov model can have only one variable. In 2003, Kersting proposed a logic hidden Markov model (LOHMM) [48] by introducing first-order logic into the Markov model. The differences from conventional Markov models are that there are two variables for the state and observation and the parameters have occurrence probabilities for each observation in addition to the transition probabilities.

(4)SRL based on random grammar

SRL [49] was proposed by Muggleton in 1996 as an extension of SLP by attaching a probability value to each clause in a random context-free grammar. In recent years, there have been many studies on SLP, such as a two-stage algorithm that learns the SLP parameters and a basic logic program proposed by Muggleton in 2000 to maximize the Bayesian posterior probability. In 2001, Cussens proposed Failure-Adjusted Maximization (FAM), which can estimate the SLP parameters on the premise that the basic logic program is given. In 2002, Muggleton proposed a method that can learn parameters and structure simultaneously. Another method based on random grammar is Programming In Statistical Modeling (PRISM) [50]. PRISM is a symbolic statistical modelling language that not only extends the probability of logic programs but also studies from examples using EM algorithms. In 1995, Sato introduced the rationale for the PRISM program through distributed semantics. In 1997, Sato and Kameya described the PRISM language and system and also gave an example to describe a hidden Markov model and Bayesian network [50]. In recent years, the study of PRISM includes a general method given by Sato in 2005. This general method first uses PRISM to write field-related logic programs, then applies the EM algorithm to learn. In 2005, Sato and Kameya proposed a new learning model based on FAM and a program conversion technique proposed by Cussens. To improve the efficiency of learning, this model adds some restrictions to the PRISM program.

1.5.2Problem introduction

The above relational learning algorithms cannot solve the learning problem of dynamic fuzzy systems. The study of relational learning has been further developed with the introduction of fuzzy theory. For example, relational learning has been enriched by fuzzy predicates [51] and a fuzzy learning algorithm called Autonomic Learning (AL) has been proposed based on fuzzy reasoning rules [52]. AL has two shortcomings: it does not consider possible noise interference in the observed data and is limited because it only considers the influence of the n th step on the n + 1th step. To solve these problems, a dynamic fuzzy relational learning (DFRL) algorithm is presented in this section.

Consider the following dynamic fuzzy inference formula:

Dynamic fuzzy rules:

If(A1,A1)and(A2,A2)and...and(AN,AN)then(B,B)Infact(A1,A1)and(A2,A2)and...and(AN,AN)Conclusion(B,B)=(A,A)(R,R)((A,A),(B,B))

where

(A,A)=(A1,A1)(A2,A2)...(An,An),(A,A)=(A1,A1)(A2,A2)...(An,An),

(A,A),(A,A),(B,B) are the DFSs on the universe (U,U),(U,U),(V,V), respectively. (R,R)((A,A),(B,B)) defines the dynamic fuzzy relation between (A,A)and(B,B), abbreviated as (R,R)

In a complex objective world, it is often difficult to determine (̄R, ̄R) based on existing fuzzy inference rules. We propose a DFRL algorithm to learn (̄R, ̄R) samples or other data obtained by means such as expert experience or monitoring system behaviour.

Given the observed data (a,a)of(A,A),(b,b)of(B,B) we have:

or

Wherej=1,2,...,M,(a,a)[(0,0),(1,1)]N,(b,b)[(0,0),(1,1)]M,(R,R)[(0,0),(1,1)]N×[(0,0),(1,1)]M.

The above can also be expressed as

where (bj,bj) is regarded as a result of applying some nonlinear transformation (rj,rj)to(a,a).

The fuzzy relational learning algorithm model proposed in [50] is as follows:

where δθ(n) is a correctional term and n, n + 1 represent successive steps of learning. The correctional term depends on the loss function lx, θ). Let the input vector be a=[a1,a2,,aN] and the expected output vector be b=[b1,b2,,bM]. Thus, x˜=[a,b],θ=[rij]N×Mi=1,2,,N,j=1,2,M,θ is defined as an element of all fuzzy relationships R in vector form. ̄c is a positive matrix of size NM × NM, which usually is a unit matrix. ∈ is a small positive number. ̄d(xk,) is a search direction vector computed from ̄x and θ.

The above model is very limited because it predicts the learning of step n + 1 from step n alone. Hence, it cannot effectively guarantee the correctness and convergence of learning. In this section, we present a solution that takes the previous k steps of learning into account and effectively deals with noise. The value of k can be determined according to the scale of the learning and the requirements of the system.

1.5.3DFRL algorithm

From the previous analysis, we should adjust the learning recursive formula and loss function of model (1.34). Thus, we can define the follow learning recursive formula:

where (wi,wi) is a coefficient representing the influence level of step i on step n + 1; its value depends on the error generated in step i. The smaller the coefficient, the greater the error, and the coefficient should satisfy i=nk+1nn(wi,wi)=(1,1). Assuming that the error in the step i learning is (ei,ei), we have (wi,wi)=[i=nk+1n(ei,ei)](ei,ei)(k1)i=nk+1n(ei,ei). The value of k (1 ≤ kn)) can be determined according to the scale of the learning and the requirements of the system. (θ,θ)=[(rij,rij)]N×Mi=1,2,,N,j=1,2,,M.(c,c) is a positive matrix of size NM × NM, which is usually a unit matrix. (ε,ε) is a small positive number. Other variables and functions are defined as in (1.34).

When the observed data contain noise, the model can be expressed as follows:

{(a,a)=(a,a)*+n(a,a)(b,b)=(b,b)*+n(b,b),

where (a,a),(b,b) are the true values of observation data (a,a),(b,b), respectively. n(a,a),n(b,b) are the noise components contained in the observed data, respectively. The variances are (δ,δ)(a,a),(δ,δ)(b,b), respectively. Thus, we can define the following loss function:

where (a,a) is an estimate of the true value (a,a).

(lj,lj)((bj,bj),(rj,rj)(a,a))=(rij,rij){12[(maxNi=1(ai,ai)(rij,rij)(bj,bj)(δ,δ)(bj,bj))2+((ai,ai)(a,a)(δ,δ)(a,a))2]}=12(maxNi=1(ai,ai)(rij,rij)(bj,bj)(δ,δ)(bj,bj))2(rij,rij)maxNi=1(ai,ai)(rij,rij)(1.39)

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
18.226.104.27