In order to be able to do classification through CART algorithm, the test data has been selected randomly from the MS dataset as 33.33%.

Let us now think for a while about how to classify the MS dataset provided in Table 2.10.1 in accordance with the steps of CART algorithm (Figure 6.29.)

Steps (13) Let us identify 66.66% of the MS dataset as the training data (D=203 × 112) and 33.33% as the test data (T=101 × 112).

Steps (410) As a result of training the training set with CART algorithm, the following results have been obtained: number of nodes –31.

Steps (1112) The splitting criterion labels the node.

The training of the MS dataset is done by applying CART algorithm (Steps 1–12 as specified in Figure 6.29) . Through this, the decision tree consisting of 39 nodes (the learned tree (N = 39)) has been obtained. The nodes chosen via the CART learned tree nodes (EDSS, MRI 1, first region lesion size) and some of the rules pertaining to the nodes are shown in Figure 6.30.

Figure 6.30: Sample decision tree structure for the MS dataset based on CART algorithm. CART algorithm graph for the sample MS dataset. (b) Sample decision tree rule space for the MS dataset.

Based on Figure 6.30, some of the decision rules obtained from the CART decision tree obtained from MS dataset can be seen as follows.

Figure 6.30(a) and (b) shows MS dataset sample decision tree structure based on CART, obtained as such; IF-THEN rules are as follows:

Rule 1

IF (first region lesion count (number of lesion) ≥ 6.5)

THEN class = PPMS

As a result, the data in the MS dataset with 33.33% portion allocated for the test procedure and classified as RRMS, SPMS, PPMS and healthy have been classified with an accuracy rate of 78.21% based on CART algorithm.

6.5.1.3CART algorithm for the analysis of mental functions

As listed in Table 2.16, the WAIS-R dataset has data of 200 people belonging to patient group and 200 sample to healthy control group. The attributes of the control group are data regarding school education, gender, …, DM. Data consist of a total of 21 attributes. It is known that using these attributes of 400 individuals, we can know whether the data patient or healthy group. How can we make the classification as to which individual belongs to which patient or healthy group and those diagnosed with WAIS-R test (based on the school education, gender, …, DM)? D matrix has a dimension of 400 × 21. This means D matrix includes the WAIS-R dataset of 400 individuals along with their 21 attributes (see Table 2.16) for the WAIS-R dataset. For the classification of D matrix CART algorithm, the first step training procedure is to be employed. For the training procedure, 66.66% of the D matrix can be split for the training dataset (267 × 21) and 33.33% as the test dataset (133 × 21). Following the classification of the training dataset being trained with CART algorithm, we can classify the test dataset.

Although the procedural steps of the algorithm may seem complicated at first glance, the only thing you have to do is to concentrate on the steps and grasp them. For this, let us have a close look at the steps provided in Figure 6.31.

How can we classify the WAIS-R dataset in Table 2.16.1 in accordance with the steps of CART algorithm (Figure 6.31)?

Steps (13) Let us identify the 66.66% of WAIS-R dataset as the training data (D=267 × 21) and 33.33% as the test data (T= 133 × 21).

2
Figure 6.31: CART algorithm for the analysis of WAIS-R data set.

Steps (410) As a result of training the training set (D = 267 × 21) with random forest algorithm, the following results have been obtained: number of nodes –30. Sample decision tree structure is shown in Figure 6.32.

Steps (1112) Gini index for the attribute is chosen as the splitting subset.

The training of the WAIS-R dataset is done by applying CART algorithm (Steps 1–12 as specified in Figure 6.31). Through this, the decision tree consisting of 20 nodes (the learned tree (N = 20)) has been obtained. The nodes chosen via the CART learned tree nodes ((logical deduction, vocabulary) and some of the rules for the nodes are given in Figure 6.31.

Based on Figure 6.32, some of the rules obtained from the CART algorithm obtained from WAIS-R dataset can be seen as follows.

Figure 6.32: Sample decision tree structure for the WAIS-R dataset based on CART. (a) CART algorithm graph for sample WAIS-R dataset. (b) Sample decision tree rule space for the WAIS-R dataset.

Figure 6.32(a) and (b) shows WAIS-R dataset sample decision tree structure based on CART, obtained as such; ifthen rules are as follows:

Rule 1

IF (logical deduction of test score result < 0.5)

THEN class = healthy

Rule 2

IF ((logical deduction of test score result ≥ 0.5) and

(vocabulary of test score result < 5.5))

THEN class = patient

Rule 3

IF ((logical deduction of test score result ≥ 0.5) and

(vocabulary of test score result ≥ 5.5))

THEN class = healthy

As a result, the data in the WAIS-R dataset with 33.33% portion allocated for the test procedure and classified as patient and healthy individuals have been classified with an accuracy rate of 99.24% based on CART algorithm.

References

[1]Han J, Kamber M, Pei J, Data mining Concepts and Techniques. USA: The Morgan Kaufmann Series in Data Management Systems, Elsevier, 2012.

[2]Buhrman, H, De Wolf R. Complexity measures and decision tree complexity: A survey. Theoretical Computer Science, 2002, 288(1), 21–43.

[3]Du W, Zhan Z. Building decision tree classifier on private data. In Proceedings of the IEEE international conference on Privacy, security and data mining. Australian Computer Society, Inc., 14, 2002, 1–8.

[4]Buntine W, Niblett T. A further comparison of splitting rules for decision-tree induction. Machine Learning, Kluwer Academic Publishers-Plenum Publishers, 1992, 8 (1), 75–85.

[5]Chikalov I. Average Time Complexity of Decision Trees. Heidelberg: Springer. 2011.

[6]Banfield RE, Hall LO, Bowyer KW, Kegelmeyer WP. A comparison of decision tree ensemble creation techniques. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2007, 29(1), 173–180.

[7]Barros R C, De Carvalho AC, Freitas AA. Automatic design of decision-tree induction algorithms. Germany: Springer International Publishing, 2015.

[8]Quinlan, JR. Induction of decision trees. Machine Learning, Kluwer Academic Publishers, 1986, 1 (1), 81–106.

[9]Breiman L, Friedman J, Charles JS , Olshen RA. Classification and regression trees. New York: Routledge, 1984.

[10]Freitas AA. Data mining and knowledge discovery with evolutionary algorithms. Berlin Heidelberg: Springer Science & Business Media, 2013.

[11]Valiente G. Algorithms on Trees and Graphs. Berlin Heidelberg: Springer, 2002.

[12]Shannon CE. A mathematical theory of communication. ACM SIGMOBILE Mobile Computing and Communications Review, 2001, 5,1, 3–55.

[13]Kubat M. An Introduction to Machine Learning. Switzerland: Springer International Publishing, 2015.

[14]Quinlan JR. C4. 5: Programs for machine learning. Elsevier: New York, NY, USA, 1993

[15]Quinlan JR. Bagging, boosting, and C4. 5. AAAI/IAAI, 1996, 1, 725–730.

[16]Quinlan JR. Improved use of continuous attributes in C4. 5. Journal of Artificial Intelligence Research, 1996, 4, 77–90.

[17]Rokach L, Maimon O. Data mining with decision trees: Theory and applications. Singapore: World Scientific Publishing, 2014 (2nd Ed.), 77–81.

[18]Reiter JP. Using CART to generate partially synthetic public use microdata. Journal of Official Statistics, 2005, 21 (3), 441–462.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
18.118.217.168