For the original data, the minimum error criterion should be satisfied in the process of dimension reduction. Suppose the first a feature values are considered. The estimated value of the corresponding reconstruction is

(X,X)¯=i=1n(Yi,Yi)(Mi,Mi)(4.16)

The average variance is

(ea,ea)=E[((X,X)(X,X)¯)2]=l=a+1n(λi,λi).(4.17)

When the a-eigenvalues of the (X,X) autocovariance matrix become larger, the mean square error becomes smaller. Thus, it is easy to find principal component whereby the former achieves a reduced dimension.

4.3.4.2Dynamic fuzzy linear discriminant analysis (DFLDA) dimensionality reduction model

DFLDA is based on LDA. The basic idea is to find some vectors that can best distinguish the classes in the DF concept space. That is, we seek to maximize the class-space metric and minimize the intra-class distance measure. The high-dimensional data can be classified into subspaces by mapping them into low-dimensional sub-spaces. Suppose that, given a dynamic fuzzy training set S, the total number of samples is N, the total number of concept classes is C, and the total number of samples in each category is Ni,i=1,2,...,c,i=1cNi=N. The dynamic fuzzy dispersion matrix in the DF concept class is

SW=1Ni=1Cj=1N((xij,xij)(xi,xi)¯)((xij,xij)(xi,xi)¯)T.(4.18)

The inter-class dispersion matrix for the DF concept is

SB=1Nj=1CNi((xi,xi)¯(x,x)¯)((xi,xi)¯(x,x)¯)T,(4.19)

where (x,x)¯ is the dynamic fuzzy mean vector of all attributes, (xi,xi)¯ is the dynamic fuzzy mean vector of the i th attribute, and (xij,xij) is the j th sample of the i th dynamic fuzzy attribute.

Therefore, according to Fisher’s linear discriminant method, get the discriminant function:

J(W)=|WTSBW||WTSWW|.(4.20)

If SW is a nonsingular matrix, then an orthogonal matrix can be obtained that maximizes the ratio of the matrix determinants of the interdisciplinary dispersion matrix to those of the intra-class dispersal distribution matrix:

Wopt*=[w1,w2,,wp]=argmaxw|WTSBW||WTSWW|,(4.21)

where {wi = 1, 2, . . ., p} is the generalized eigenvector of (4.20), corresponding to the p largest maximum eigenvalues of (4.20), which is

SBwi=λiSwwi.(4.22)

As the rank of the inter-class scatter matrix SB of the DF concept is less than or equal to C − 1 generalized nonzero eigenvalues. W opt is the best projection transformation matrix.

According to the eigenvector corresponding to the largest eigenvalue (the number of elements in the eigenvector is the number of attributes), all the attributes in the eigenvectors with values above a certain threshold (δ,δ)((0.0)<(δ,δ)<)(1,1)) are selected as attributes of the new training samples.

4.4Concept learning model based on DF lattice

The formal concept analysis theory was first applied to the concept of discovery, display, and sorting. One of the main data structures for formal concept analysis is the concept lattice. Each node in the concept lattice is a concept. An extension refers to the instance space (object) covered by the concept. The connotation is the common characteristic (attribute) of the covered instance. The structure reflects the relationship between objects and attributes and a complete concept hierarchy of conceptual generalizations and specialization relations. Burusco and others [6] first applied fuzzy theory to formal concept analysis, and constructed a lattice structure based on the L-fuzzy concept set. As a result of the interaction between the theory of order and lattice theory, the combination of the concept lattice with DFL has a very important practical value. In this section, we introduce four classic construction algorithms; then, we focus on the hierarchical DF concept lattice construction algorithm before discussing the DF concept lattice reduction technology. Finally, we introduce a DF containing rule generation strategy and verify the effectiveness of the algorithm by means of an example and an experiment.

4.4.1Construction of classical concept lattice

At present, concept lattice construction algorithms are divided into two kinds: batch algorithms and increment algorithms. Batch-based construction algorithms are mainly applied to relatively small instance sets in the data space, e.g. the Bordat, Chein, Ganter, and Nourine algorithms. Incremental construction algorithms are applied to instance spaces where the data are constantly updated. Representative algorithms include those of Godin, Capineto, Ho, and so on.

4.4.1.1Batch algorithms

Batch processing algorithms to generate a concept lattice generally have two steps: First, the grid nodes are generated; second, the grid nodes are used to establish a direct precursor/direct successor relationship. According to the task order, there are two main generation methods, namely, the task partitioning model (as developed by Chein, Ganter and other algorithms) and the task cross-building model (such as Bordat’s algorithm).

The Chein algorithm [7] constructs the concept lattice from bottom to top. The algorithm iteratively generates a collection of all concepts starting with the most ubiquitous, such as the concept of an object. The algorithm starts at the L1 level and contains a set of pairs ({x}, f(x)) for all objects x in all object sets. It constructs a new set of pairs of Lk + 1 layers by combining any pair of Lk layers. The merging process considers the two pairs of intersection sets for the Lk layer and tests whether the intersection already exists, depending on the result. If the upper layer appears, it indicates that the upper layer is not completely correct and is marked as such, so that the end of this layer operation will be deleted. The algorithm is simple and clear, but the practical application will produce a large number of redundant pairs in the same layer and in lower layers. In the algorithm, the same concept will be generated at different times by different concepts. This is the main drawback of this algorithm. The algorithm itself does not generate a Hasse graph, but this layer-by-layer grid-node generation method is useful for identifying links between grid nodes.

In the Bordat algorithm [8], the concept lattice L is constructed from the topmost node. For each node, the child nodes are generated and the link between child nodes and parent node is completed. The process is then called recursively for the resulting child nodes. The basic idea of the algorithm is that if the current node is (O1, D1), where O is the object set and D is the attribute set, we find the attribute subset D2 = D such that D2 can maintain the complete binary property on O1. This is the largest expansion and constitutes the current node of the subnode’s content. The algorithm is also very simple, intuitive, and easy to run in parallel. The disadvantage is that the same node is generated multiple times. In practical applications, each node is duplicated to generate its parent node several times. However, this is one of the most effective algorithms.

The Ganter algorithm [9] uses the eigenvectors representing X sets to enumerate X sets of cells. The length of each vector is the cardinality of the attribute set. The corresponding bit is set to 1 if an attribute value is present in the vector, and to 0 otherwise. The feature vectors are sorted in lexicographic order. Given a vector X = {x1, x2, ..., x m}, it finds the next X vector. The way to do this is to set the property bit in order and test whether it is a complete pair. The algorithm is initialized with {0, 0, ..., 0}. The vectors produced in this way are ordered. The list is sorted according to the containing topology of the X set. The algorithm does not generate Hasse graphs.

In contrast to the other algorithms, the Nourine [10] algorithm generates a concept lattice using the concept of epitaxial complement as the basic operating units. The algorithm is divided into two parts. First, the extension set of all the individual attributes forms a base B, and the set of clusters generated in B is F={DKU|KB}. The purpose of this step is to generate F from the base as a lexical-order tree representation. Each path of the tree represents an element in F. Operations with the same prefix share the same partial path. The root is an empty set. The overlay of F is then calculated. The covering relation between two elements is defined in the algorithm, and the condition whereby the elements in the lexical-order tree satisfy the covering relation is derived to determine the corresponding algorithm. It has been shown that the algorithm is less complex than those of Bordat and Ganter et al.

4.4.1.2Incremental algorithms

The basic idea of increment algorithms is to present the object to be inserted and place all the concepts in the intersection of the lattice according to the results of different treatments. It is assumed that the first i objects in the formal context have generated subsets of the concept nodes and sublattices of the concept lattice. This step is repeated until the final cell structure is generated. Obviously, it is easy to apply this strategy to attributes, which can be generated incrementally based on attributes rather than objects. Typical algorithms include Godin’s algorithm [11], Capineto’s algorithm [12], and Ho’s algorithm [13]. The property-based incremental construction algorithm is similar to the basic idea of Godin’s algorithm. The difference is that the algorithm adds to the concept lattice one by one.

Given the concept lattice L corresponding to the background B: (X, Y, I) and the new object X, the concept lattice L corresponding to the formal context B = (X ∪{X}, Y, I) is given. There are three main problems in the process of generating the concept lattice: (1) generation of all new nodes, (2) avoiding the repetition of existing lattice nodes, and (3) updating the edges. Therefore, for each node in the original concept lattice, different types are defined according to the relation between them and the connotation description of the new object.

Definition 4.20 If a lattice node C satisfies Internt(C) ⊆ f({x}), C is called a replacement node. Obviously, if C is a replacement node, then C should be replaced by Extent(C) ⊆ {X}, Intent(C)) in L.

Definition 4.21 If the given node Ci = (Oi, Ai) satisfies the following: (1) the intersec tion of connotations (f({x}) between the concept node and new node is not equal to the connotation of any node in the lattice; (2) the intersection of connotations (f({x}) between the concept node and new node is also not equal to Intent(C)∩Cj(f({x}), where Cj is any node that satisfies Cj > C in L; then Ci is called an extended subgrid node.

Obviously, replaceable nodes cannot be generated with the same node because Dif({x*})=Di. If a node in lattice L is neither a replacement node nor a generatable node, it is called an invariant node.

For any node C in L: If there is no node C in L satisfying Intent(C) = Intent(C), C is called a new node; otherwise, it is called an inheritance node. If C = is an extended ′ subgrid node, node Extent(C), Internt(C) ∪ {X}) is obviously a new node in L. This is called a newborn node generated by C′.

Theorem 4.4 Given the concept lattice L and the new object x, the node C in L is a generating subgrid node if and only if (1) ¬(Internt(c1)f{X*})and(2)¬(C2)BeforeC1IntentC2 = IntentC1, where Before(C1) denotes the set of all lattice nodes that are accessed before C1 and all the newborn lattices generated before C1. The order of access here is given by the L grid nodes in accordance with the meaning of small to large access to the order. This ensures that the C1 predecessor node is accessed first to satisfy the definition of generating the child nodes.

Theorem 4.5 If C1 is a subgrid node and its corresponding newborn node is Cnew, then L satisfies the concept of Intent(C2) ⊇ Intent(Cnew), and C2 must satisfy Intent(C2) ⊇ Intent(C1).

It can be seen from Theorem 4.5 that, for a new node, none of the old nodes can be its subnode except for its subgrid node.

Theorem 4.6 For two new nodes Cnew1 and Cnew2, if Intent(Cnew1) ⊂ Intent(Cnew2), then Cnew1 must be generated before Cnew2 and cannot become its subgrid node.

4.4.2Constructing lattice algorithm based on DFS

4.4.2.1Delamination of DF concept lattice

In the concept lattice of DF, the hierarchical structure of a concept is defined by attributes. Therefore, the hierarchical concept lattice is defined by the length of the attribute set. We now introduce the DF concept lattice.

Definition 4.22 In the concept lattice of DF, if the longest distance between the concept of ((0,0),(A,A)) and the largest element in the lattice is N, then we say that DF concept ((0.0),(A,A)) is in the Nth layer of the DF concept lattice.

Definition 4.23 In the DF concept lattice, we assume that ((0i,0i),(Ai,Ai))((0j,))0),(Aj,Aj)) and if the longest chain length of ((0i0i),(Ai,Ai))to((oj,0j)),((Ai,Ai))isl(lZ+),then((0i,0i),(Ai,Ai)) is the l-layer ancestor of ((0j,0j)),(Aj,Aj)).

Obviously, there is no overlay between two nodes on the same layer and there is no connection line in the DF concept lattice. The high-level nodes at different layers are covered by at least one lower node.

Definition 4.24 If concepts (ci,ci)and(cj,cj)satisfy(ci,ci)(cj,cj), then (ci,ci) is a subconcept of (cj,cj),and(cj,cj) is a superconcept of (cj,cj).

If there is no concept (ck,ck) in the concept lattice such that (ci,ci)(ck,ck)(cj,cj) holds, then we call (cj,cj) the parent node of (ci,ci) and (ci,ci) is the child node of (cj,cj) There is an edge between (ci,ci) and

Definition 4.25 Except for the minimum element, the DF concept lattice can be divided into Tmax=(x,x)max(x,x){|(x,x)|} layers, among which |(x,x)| is the number of attributes of the object.

4.4.2.2Algorithm for constructing the DF concept lattice layer by layer

Construction of the DF lattice algorithm from top to bottom (DFCL_Layer) is as follows:

All of the temporary children of the second layer are generated iteratively according to this algorithm.

Processing operations on temporary subnodes: The (N + 1)th temporary subgrid node is ((0N+1,1,0N+1,1),(AN+1,1,AN+1,1))((0N+1,2,0N+1,12),(AN+1,2,AN+1,2)),......((oN+1,q,oN+1,q),(AN+1,q,AN+1,q)) Sort these according to the number of objects covered by the node, i.e. ......|(oN+1,i,oN+1,i)|(oN+1,i,oN+1,i)|......, where 1i,jq,and,ij Finally, the neighbouring temporary sublattice nodes are examined in turn. If (0N+1,i,0N+1,i)(0N+1,j,0N+1,j), then delete the temporary subnode <(0N+1,i,0N+1,i),(AN+1,i,AN+1,i)> and judge the sorted temporary subgrid nodes until all subgrid nodes of the (N + 1)th layer have been determined.

The algorithm is simple and clear and is fully dependent on the dynamic fuzzy processing background. The construction of a dynamic fuzzy concept lattice is faster than in other algorithms.

4.4.2.3A critical hierarchical construction algorithm for DF concept lattice

The DF concept lattice critical layer construction algorithm (DFCL_BorderLayer) is the first to find all critical concepts. According to a list of properties, the algorithm then constructs the DF concept lattice. The definition of the critical concept is realized by the intersection of the connotation and the union of the extension of the concept of DF. The algorithm is a bottom-up hierarchical construction algorithm.

Algorithm 4.1 Constructing the DF concept lattice layer by layer

First, according to the prior knowledge, we give the appropriate threshold to preprocess the dynamic fuzzy formalism background (B: (X, Y, I)). Then, we remove the meta-sets that satisfy the following two conditions.

(1){(x,x)|(x,x)X,sothat,f((x,x))=Y} which meets all the attributes of the sample collection.

(2){(y,y)|(y,y)y,sothat,g((y,y))=x}, which covers all instances of the property set.

1)Assuming that the first j-1 lattice <(0N,1,0N,1),(AN,1,AN,1)>,...,<(0N,j1,0N,j1),(AN,j1,)AN,j1)> temporary grid nodes of the Nth layer have been generated, the following will generate the temporary grid nodes of <(0Nj,0Nj),(ANj,ANj)>.LetM=Yj=1r(ANj,ANj).ifMϕ, then take (yi0,yio) to make (g((ANj,ANj)(yj0,yjo)),(x,x)eg((ANi,ANj))(yj0.yj0)f((x,x))) Otherwise, go to (9).

2)Generate denoted as (αj0,βj0).

3)Set ℘ = {αj0}, and let i = 0; O = (αj0} (O is a collection of object sets).

4)LetM=Y(j=1r(ANr,ANr)βij).ifMϕ,take(yj,i+1,yj,i+1)make|g((AN,j,AN,j))(yj)(yi,(j+1),)|=(y,y)emMax{|g(AN,iAN,i),(y,y))|},andget(g(AN,i,AN,i)(yi,(i+1),yj,(j+1))),(x,x)eg((AN,j.AN,j)(yj,(j+1).yj.(i+1)))f((x,x))),denotedas (αj,i+1,βj,i+1);otherwise,gotostep(9).

5)Set αj,i+1(i.e.αi,j+1),gotostep(5)

6)If there is an element in ℘ that contains (αi,j+1,βi,j+1)to(αi,j+1,βi,j+1). go to step (5).

7)Produce and change ={αi,j+1},i++;

8)Reset

9)If M = 0, terminate the algorithm.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
18.191.192.59