4.3. Reasoning (Inference) Networks (1)

A great number of reasoning processes can be represented by a graph search.
First, let see an example.

Example 4.1

The jealous husbands problem is one of the well-known river-crossing puzzles. Three married couples denoted by {A a, B b, C c} must cross a river using a boat which can hold at most two people and subject to the constraints: (1) no woman can be in the presence of another man unless her husband is also present, (2) only three men {A,B,C} and one woman {a} master rowing technology. Under the constraints, we have two heuristic rules: (1) there cannot be both women and men present on a bank with women outnumbering men, since if there were, some woman would be husbandless, (2) only the following ways of river crossing exist: one or two men, woman a or with any other woman, or any couple. Where the married couples are represented by A (male) and a (female), B (male) and b (female), and C (male) and c (female), respectively.
Using the heuristics, we have a shortest solution to the problem shown in Table 4.1. There are 13 one-way trips. The river-crossing process can be represented by a graph as shown in Fig. 4.1.
That is, given a directed graph G, a premise A and a goal p in G, a reasoning process can be regarded as finding a path from A to p in the graph. Since a directed acyclic graph can be represented by a semi-order space image, the reasoning mechanism can then be depicted as follows.
(1) image is a premise, Let image and a goal image
(2) Reasoning rule: if b is a direct successor of a, then image.
If we find a p in image such that image, then p holds. Otherwise, p does not hold.
In order to change the certain reasoning model above into an uncertain reasoning model, the following two problems must be solved: (l) the representation of the uncertain premise; (2) the representation of the uncertain reasoning rule. As mentioned before, uncertain information can be described in different grain-size spaces, while uncertain reasoning rules can be represented by probability. We have the following uncertain reasoning model:
(1) For image, we define a function image on A
(2) D is a set of edges on a reasoning network. Define a reasoning function image
(3) Reasoning rule: if b is a direct successor of a, for image, define an reasoning rule image
F is an arbitrary combination function, for example, image
(4) Given a goal image, by the model we have image, then p holds with confidence d.

Table 4.1

River-Crossing Process

Trip NumberStarting BankTravelEnding Bank
StartA a B b C C
1A B C ca b →
2A B C c← ab
3A B Ca c →b
4A B C← ab c
5A aB C →b c
6A a← B bC c
7B bA a →C c
8B b← C cA a
9b cB C →A a
10b c← aA B C
11ba c →A B C
12b← aA B C c
13a b→A B C c
FinishA a B b C c

image

image
Figure 4.1 River-Crossing Graphical Representation
Especially, when d = 1, p holds. Otherwise, if d = 0, p does not hold.
Meantime, another two problems must be solved in order that reasoning can be made in a multi-granular world. First, given an uncertain reasoning model on X and its quotient space [X], how to construct a corresponding reasoning model in [X]? This is a projection problem. How to obtain some useful information on X from the reasoning that we draw on [X], this is the problem about the relation between two reasoning models. Second, given two reasoning models on quotient sets image and image, how to get a whole understanding about X? This is the synthesis problem of different evidences. We first discuss the projection problem.

4.3.1. Projection

An uncertain reasoning model is represented by

image

Where X is a domain or a set of nodes of graph G. D is a set of edges in G. A is a subset of premises, p is a goal, image is a reasoning function defined on A. T is a semi-order structure. F: image is a function called a reasoning rule.
In essence, so-called reasoning is to answer whether p can be inferred from A, or to what degree of confidence p does hold, when given domain structure T and reasoning rule F.
Projection problem is that given reasoning model with respect to X and its quotient set X1, find the corresponding reasoning model on X1.
Since X1 is a quotient set of X, Edges image on X1 is the corresponding quotient set of D. First, we define

image

Let image be a set of all equivalence classes image defined above. At present image is not a partition of D. Let E0 be a set of edges of D not belonging to any equivalence class in image. Adding E0 into image, we have D1. D1 is a quotient set of D. image is undefined in E0.
A1 is a quotient set with respect to A. p1 is the equivalence class corresponding to p, and image.
R is an equivalence relation with respect to X1. image on image is an induced quotient quasi-semi order from R. If R and T are compatible, then image is a quotient semi-order on X1.
Again, the methods for constructing f1 and g1 are given below. For example, by using ‘combination principle’, we define

image

image

where, G1 and G2 are arbitrary combination functions of their arguments. The range of g1 is image.
Taking the same F as the reasoning rule, we have a reasoning model of X1 as

image

Example 4.2

Assume that

image(4.1)

image

image(4.2)

We have the following proposition.

Proposition 4.1

image is known. After reasoning, we have image. Let image be the corresponding reasoning model on image, where image is a quotient set of X. image and image are defined by Formulas (4.1) and (4.2). Then from the reasoning model on image, we have image.

Proof

Assume that in X from image and via image, we infer p and image.
According to the reasoning steps, by induction, when image, the conclusion obviously holds. Now assuming that for n < m, the conclusion holds, we show that for n = m the conclusion still holds.
Assume that image. From the assumption, we have image.
(1) If image, since image, we have

image

Since image, we have image. Namely

image

(2) If image, assuming that image, since image, we have image. From the above definition, we have

image

Again,

image

image

The proposition indicates that the projection defined above satisfies the ‘homomorphism principle’. That is, if in the high (coarse) abstraction level, we infer that p1 holds with confidence <d, then in the low (fine) level, the corresponding p holds with confidence <d as well.
Simply speaking, if there is no solution in some regions of high abstraction level, i.e., p1 holds with low confidence, then, there isn’t any solution in the corresponding regions of any lower level. This means that based on the result that we infer from high abstraction levels, the possible solution regions in low levels are narrowed. This is the key for reducing computational complexity by hierarchical problem-solving strategy, as discussed in Chapter 2.
When equivalence relation R corresponding to X1 is incompatible with the semi-order structure T of X, we must revise R such that it is compatible with T, as we have discussed in Chapter 1.

4.3.2. Synthesis

In our reasoning model, uncertain information is represented at different grain-size worlds. The information observed from different perspectives and aspects will have different forms of uncertainty. We will consider how to synthesize the information with different uncertainties.
Assume that X1 and X2 are two quotient sets of X. Two reasoning models image and image on X1 and X2 are given. The synthesis of reasoning is to find a new reasoning model image in the synthetic space X3 of X1 and X2.
According to the approach indicated in Chapter 3, we have
(1) X3: the least upper bound of spaces image and image
(2) D3: the least upper bound of spaces image and image
(3) image: the least upper bound of semi-order structures image and image.
In fact, when X3 and image are fixed, D3 is uniquely defined. Therefore, X3 is defined first, then image and finally D3 is obtained from image.
After the projection operation of attribute functions on X has been decided, the synthetic functions image and image are known based on the synthetic method shown in Section 3.6. Let

image

image

If the same reasoning rule F is used, we have a reasoning model in the synthetic space X3 as follows.

image

We next use our synthetic approach to analyze Dempster-Shafer combination rule (or D-S combination rule) in Belief Theory.

Example 4.3

First, we introduce briefly some basic concepts in Belief Theory (Shafer, 1976).
Assume that X is a finite domain, image is a power set of X. Define a function image such that
(1) image
(2) image
m is called a basic probability assignment on X. Let image. Bel(A) is said to be a belief function.
Since belief function (Bel) is defined on P(X), when P(X) is regarded as a new set X1, then Bel is just a function on X1. Therefore, Bel is an attribute function on P(X).
Assume that m1 and m2 are two given attribute functions on X1 and X2, respectively. We have two problem spaces image and image, where image is a topology on image, i = 1, 2.
The synthesis of the two spaces is as follows.
Let X3 be the least upper bound of spaces X1 and X2, T3 be the least upper bound of topologies T1 and T2.
Define a projection operation of attribute functions as follows.
Assume that m3 is an attribute function on X3.
Letting image :image be a nature projection, we have image. That is,

image(4.3)

As mentioned in Section 3.6, the synthesis of quotient spaces image and image is equivalent to that m3 satisfies Formula (4.3). But the solution of m3 is not unique. Some optimal criterion must be introduced.
It’s known that the attribute function in the coarse level, except itself, cannot provide more information to the attribute function in the fine level. In other words, besides that m3 satisfies (4.3), it must preserve a maximal uncertainty, or in terms of information theory, ‘a maximal entropy’.
When X is a finite set, the entropy of image can be defined as

image(4.4)

The maximal criterion is

image(4.5)

where the range m is all basic probability assignments which satisfy (4.3).
When X is a finite set, the two quotient spaces of X are

image

The least upper bound space X3 of X1 and X2 is

image

We first assume that all image. Let

image

From (4.3), we have

image(4.6)

From (4.4), we can write

image(4.7)

Under the constraint (4.6), we next find the maximum of Formula (4.7).
Using Lagrange multiplier, we have

image

Finding the partial derivative of the formula above with respect to each image, and letting its result be zero, we have a set of equations

image(4.8)

Letting image and image, substituting into (4.8), we have

image(4.9)

Again from (4.6), we can write

image(4.10)

Since m1 and m2 are basic probability assignments in image and image, we have

image

Letting image, we obtain image, image and t-s=0.
Letting t=s=1, we have

image

Finally, image.
This is just the D-S combination rule, where image, image and image are basic probability assignments. The combination rule that we obtained is from the maximal entropy principle.
In case of some of image being empty, we will discuss this next.

Example 4.4

image is a domain. image and image are two quotient sets of X. Let

image

image

Since image, image, image and image, we have

image

If image satisfying Formula (4.6) exists, from image, we have

image

Since image, image.
That is, image. But from image, we have image. This is a contradiction. Therefore, there does not exist such a image that satisfies Formula (4.6).
As discussed in Section 3.6, when the image satisfying Formula (4.6) does not exist, similar to the least-squares, a criterion can be introduced. Combining with the given criterion, an optimal solution can be obtained.
When some of image are empty, image is defined as follows.
Let

image

image

image

Formula (4.6) becomes

image(4.11)

Formula (4.11) may not have a solution, so a weighted least-squares function is used as shown below.

image

where image and image are weights (constants).
As an optimal criterion, ‘the maximal entropy principle’ is still used. That is,

image

Let

image(4.12)

Find a image such that the right hand side of Formula (4.12) is minimum.
Let image and subject to the constraint

image(4.13)

By using the Lagrange multiplier, we have

image(4.14)

Finding the partial derivative of Formula (4.14) with respect to each image and letting the result be zero, we have a set of equations.

image(4.15)

Where, the first sum image is only over all j in image, the second sum image is only over all i in image.
For image in image, letting image and substituting into Formula (4.15), we have

image

Or

image(4.16)

Let

image(4.17)

Formula (4.16) changes as image. We have image.
Since image, we have image, i.e., image.
Finally, we have

image(4.18)

Where, the sum image is over all image.
Formula (4.18) is just the D-S combination rule.
We therefore conclude that, the D-S combination rule can be inferred from the synthetic method of quotient spaces by using ‘maximal entropy principle’. It is noted that D-S rule is simple and is easy to be used, but weights image and image shown in Formula (4.17) are artificial. It means that we might choose some other optimal criteria such that different combination rules would be obtained to aim at different issues.
Domain structural knowledge, the relationships among elements, for example, the age order among people, the Euclidean metric relations in a two-dimensional image, etc., is not considered either in membership function or in belief function. But it is quite important in multi-granular computing. In our model, taking all factors, including domain, structure and attribute, into consideration and when being transformed from one grained world to the others, their structures are required to satisfy the homomorphism principle. By the principle, it’s known that if a proposition is rejected in a coarse-grained world, it must be false in the fine-grained world. Therefore, due to the principle we can benefit from multi-granular computing.

4.3.3. Experimental Results

In this section, we will show the advantage of multi-level graph search (or reasoning) based on quotient space theory by data validation (He, 2011).
The experimental setting is searching the shortest path on three different kinds of networks, i.e., random, small-world and scale-free networks that randomly generated from data. The numbers of their nodes are 100, 200, 300, 400 and 500, respectively. The experiments are implemented on a Pentium 4 computer (main frequency 3.00 GHz, memory capacity 2 GB) using Windows XP SP3. Our method is multi-level (generally, 5–6 levels) search based on quotient space theory. Dijkstra and Floyd algorithms are well-known search algorithms (see Dijkstra, 1959; Floyd, 1962, for more details).
From the results (Tables 4.24.4), it’s shown that two orders of magnitudes of CPU time can be saved by multi-level search when the scale of networks becomes larger. But the performances slightly reduce, i.e., the ratio between the optimal path found and the real shortest path is 85% in the multi-level search.

Table 4.2

Total CPU Time (in seconds) in Random Networks

Number of Nodes100200300400500
Our method0.9391.3563.1126.63412.171
Dijkstra1.7199.79791.141565.3911002.125
Floyd0.9404.220118.630212.030511.560

image

Table 4.3

Total CPU Time (in seconds) in Small-World Networks

Number of Nodes100200300400500
Our method0.6402.4036.65914.33022.262
Dijkstra2.75810.54687.239700.5401100.200
Floyd0.7904.783132.743230.412498.317

image

Table 4.4

Total CPU Time (in seconds) in Scale-Free Networks

Number of Nodes100200300400500
Our method0.4371.7344.2978,18715.562
Dijkstra2.0458.23686.431668.400998.354
Floyd0.7804.060108.598206.580466.250

image

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
18.118.142.250