a knowledge roughness based approach to hybrid decision tree construction

5
Procedia Engineering 29 (2012) 1544 – 1548 1877-7058 © 2011 Published by Elsevier Ltd. doi:10.1016/j.proeng.2012.01.170 Available online at www.sciencedirect.com 2012 International Workshop on Information and Electronics Engineering (IWIEE) A Knowledge Roughness Based Approach to Hybrid Decision Tree Construction Hongmei Lu a * ,b , Xuegang Hu b , Baosheng Yang a a Laboratory of Intelligent Information Processing, Suzhou University, Suzhou 234000, China b School of Computer Science and Information EngineeringHefei University of technology, Hefei 230009, China Abstract It is difficult for unvaried decision tree to reflect the relationship of attributes, multivariate decision tree can resolve this problem preferably, the former produces big tree, the latter gains simple tree but difficult to explain. Aim to these two features for decision tree, in this paper we advance a knowledge roughness based approach to hybrid decision tree, select lesser knowledge roughness as tested attribute to construct decision tree. As a result ,we find this is a good approach with simple operation and higher efficiency. © 2011 Published by Elsevier Ltd. Keywords: Rough sets; Knowledge roughness; Univariate decision tree; Multivariate decision tree; Hybrid decision tree 1. Introduction Decision tree classification method has been widely used for its advantages of high speed, high precision and simple generating model [1] .Most decision tree construction method only test single attribute on each node, this single-variable decision tree ignores the properties association between the information system, and often costly in trimming [2] . According to the concept of core(condition attributes related to decision attributes in the rough set theory) and the relatively generalized structure, Duo-qian Miao and others have tested multivariate, proposed a multi-variable decision tree construction method [3] ,the method can construct a very simple decision tree, but when the core has more attributes ,the node has too much properties to descript the node splitting conditions, and difficult to understand the constructed decision tree. Studies have shown that the complexity of a single variable decision tree mainly decided by the * Corresponding author. Tel/ fax: +86-557-287-1028. E-mail address: [email protected].

Upload: hongmei-lu

Post on 29-Nov-2016

212 views

Category:

Documents


0 download

TRANSCRIPT

Procedia Engineering 29 (2012) 1544 – 1548

1877-7058 © 2011 Published by Elsevier Ltd.doi:10.1016/j.proeng.2012.01.170

Available online at www.sciencedirect.comAvailable online at www.sciencedirect.com

Procedia Engineering 00 (2011) 000–000

ProcediaEngineering

www.elsevier.com/locate/procedia

2012 International Workshop on Information and Electronics Engineering (IWIEE)

A Knowledge Roughness Based Approach to Hybrid Decision Tree Construction

Hongmei Lua*,b, Xuegang Hub, Baosheng Yanga

aLaboratory of Intelligent Information Processing, Suzhou University, Suzhou 234000, China bSchool of Computer Science and Information Engineering,Hefei University of technology, Hefei 230009, China

Abstract

It is difficult for unvaried decision tree to reflect the relationship of attributes, multivariate decision tree can resolve this problem preferably, the former produces big tree, the latter gains simple tree but difficult to explain. Aim to these two features for decision tree, in this paper we advance a knowledge roughness based approach to hybrid decision tree, select lesser knowledge roughness as tested attribute to construct decision tree. As a result ,we find this is a good approach with simple operation and higher efficiency.

© 2011 Published by Elsevier Ltd.

Keywords: Rough sets; Knowledge roughness; Univariate decision tree; Multivariate decision tree; Hybrid decision tree

1. Introduction

Decision tree classification method has been widely used for its advantages of high speed, high precision and simple generating model[1].Most decision tree construction method only test single attribute on each node, this single-variable decision tree ignores the properties association between the information system, and often costly in trimming[2]. According to the concept of core(condition attributes related to decision attributes in the rough set theory) and the relatively generalized structure, Duo-qian Miao and others have tested multivariate, proposed a multi-variable decision tree construction method[3],the method can construct a very simple decision tree, but when the core has more attributes ,the node has too much properties to descript the node splitting conditions, and difficult to understand the constructed decision tree. Studies have shown that the complexity of a single variable decision tree mainly decided by the

* Corresponding author. Tel/ fax: +86-557-287-1028. E-mail address: [email protected].

1545Hongmei Lu et al. / Procedia Engineering 29 (2012) 1544 – 15482 H. M. Lu et al./ Procedia Engineering 00 (2011) 000–000

number of decision tree nodes, while the complexity of a multi-variable decision tree mainly decided by the number of attributes of decision tree nodes[4]. For the above two kinds of tree characteristics, Hu Xue-gang has put forward a mixed variable decision tree structure[5],upon which we propose a knowledge roughness based approach to hybrid decision tree construction, selecting the attribute with smaller knowledge roughness to construct a decision tree. The experimental results show that this is a simple, efficient method of decision tree generation.

2. Basic Concepts

Definition 1 For a knowledge representation system, an information table S=<U,R,V,f>,whereU is a collection of objects, also known as the domain, R is the set of attributes, V is the set of attribute values, f:U×R→V is an information function. If the attribute set of R can be divided into two disjoint sets: conditional attributes sets C and decision attribute set D, to meets=C∪D and=φ , such information systems, also knows as decision-making system.

With regard to the basic concepts of rough sets, such as equivalence relations, equivalence classes, as well as the statements of upper approximation, lower approximation and boundary can be found in references[6,7].

Definition 2 suppose A⊆C,B⊆D,U/A={X1,X2,…,Xn} and U/B={Y1,Y2,…,Ym} are provided by the equivalence relation to the U of A and B division, then the subset Yi of U/B set is a rough set based on knowledge A in the domain of U, define its A precision (in the case of confusion does not occur, also called precision) are:

dA(Yi) =│A-( Yi)│/│A-( Yi)│ Yi≠φdA(Yi) = 1 Yi =φ

definition of A roughness degree(in the case of confusion does not occur, also referred to as roughness degree) are:

PA(Yi)=1-dA(Yi)Definition3 defines Min(PA(Yi),(i=1,…,m))as roughness degree PA(B)of knowledge A

for a collection of U/B.Definition 4 set P and Q are two equivalence relation families for U, and: U/IND(P) = {X1,

X2,…,Xn}, U/IND(Q) = {Y1,Y2,…,Ym}, so that

U

U

K

)(1

)(

},:{

),,2,1(

,}:{

PINDXijjm

PINDXijji

j

j

iYXXZ

mi

YXXZ

∈+

∀⊄=

=

⊂=

We call the equivalence relation {Z1,Z2,…,Zm+1} in U as the generalization of P relative to Q[3],record as GENQ(P)。

With the equivalence relation generalization to instead of the conjunction of original property values as a multi-variable testing can avoid overfitting and also simplify the decision tree.

Definition 5 Decision tree learning algorithm selects the properties as little as possible to classify more instance on each node according to concrete data set, which determines the current number of node

1546 Hongmei Lu et al. / Procedia Engineering 29 (2012) 1544 – 1548 H. M. Lu et al. / Procedia Engineering 00 (2011) 000–000 3

splitting properties, so on a decision tree, one node may be split by single-variable, while the other node split by multi-variable, then we call it a hybrid decision tree.

3. Algorithm

In this paper we propose a hybrid decision tree construction algorithm based on knowledge rough degree(KRDH),whose input is a collection of data to be classified, the output is a decision tree. The specific algorithm is as follows:

(1) in condition attribute set C, for each unselected property(assuming there are k) calculating their knowledge rough degree Pci(D), if Pci(D)=1,then transfer (5);

(2) if Pci(D)=0,then selects this single- variable attribute Ci as root node, labled it, go to (1); (3) else select the two condition attributes with minimum and sub-small value of knowledge rough

degree( if present, without selection for Pci(D)=1 )to generalize the domain U={Z1,Z2,…,Zm+1},forming multivariate splitting attributes, marking Z1,Z2,…,Zm is a leaf node;

(4) if Zm+1=φ,labled it as a leaf node, the algorithm end; else set Temp U = {the instance collection of Zm+1}, U = Temp U, go to (1); (5) for each unselected attribute in C, calculating their information entropy, selecting the attribute of

the smallest entropy as a single variable to split domain, marking the attribute has been selected, transfer (1).

4. Case analysis

We select the lenses data set from UCI international machine learning library as an example to describe the workflow of KRDH algorithm, comparing with the traditional construction ID3 decision tree algorithm based on the information gain, indicating the feasibility and superiority of our algorithm.

The lenses data set has a total of 24 records, each record corresponds to five properties, of which the first four attributes A1, A2, A3, A4 for the condition of property, the last one attribute D is decision attribute.

According to the definitions in section 2 and algorithm given in section 3, Instance constructed by the roughness of the hybrid knowledge-based decision tree variables TH, as shown in Figure 1, The figures in brackets represent each leaf node contains a number of instances. Figure 2 shows the information gain is based on the decision tree ID3 Methods TE , obviously, TH much simpler than the TE, the rules are concise than the TE.

Fig.1. mix-variable decision tree based on knowledge roughness TH

1547Hongmei Lu et al. / Procedia Engineering 29 (2012) 1544 – 15484 H. M. Lu et al./ Procedia Engineering 00 (2011) 000–000

Fig.2. decision tree based on information entropy TE

5. Experimental analysis

To facilitate comparison, we realized KRDH algorithm, under Matlab 6.5 platform and selected 5 data sets from UCI international machine learning library, compared the size of the constructed decision tree, and with cross validation compared the classification algorithm accuracy. Experimental environment is P42.0G, 512M memory.

(1) Comparison of DT size

Table 1 The leaf node / total node comparison of KRDH and ID3 algorithm.

Data set Tuples Attributes(C/D) ID3 KRDH

Monk2 432 6/1 290/442 210/336

House-votes84 435 16/1 161/285 148/276

Temp 11232 6/1 4/39 4/39

Tic-tac-toe 958 9/1 218/343 141/236

Balance 625 4/1 230/325 193/292

(2) Comparison of classification accuracy

Table 2 The classification accuracy comparison of KRDH and ID3 algorithm.

Data set Tuples Attributes(C/D) ID3 KRDH

Monk2 432 6/1 44.18% 58.83%

House-votes84 435 16/1 39.76% 55.11%

Temp 11232 6/1 49.11% 49.11%

Tic-tac-toe 958 9/1 55.78% 63.15%

Balance 625 4/1 4.83% 16.45%

1548 Hongmei Lu et al. / Procedia Engineering 29 (2012) 1544 – 1548 H. M. Lu et al. / Procedia Engineering 00 (2011) 000–000 5

Experimental data show that: compared with the traditional ID3 decision tree algorithm, whether the size of the tree or the classification accuracy, the knowledge roughness based variables decision tree algorithm has improved considerably.

6. Conclusions

Decision tree is an important form of Multi-attribute inductive learning, Unvaried decision tree stresses the fineness of the division of knowledge, neglecting the association between the properties widespread information systems, the choice of property has some flaws, it is difficult to get the best decision tree. Miao Duo-Qian et al proposed the construction method of multi-variable decision tree ,very simple decision tree can be constructed, to better reflect the relationship between attributes, but when the properties of the nucleus are large, the properties of decision tree nodes have more, Leading to difficulties in the description of splitting node, So it is difficult to understand the structure of the decision tree. The hybrid variable decision tree can better make up less than two, because the roughness of knowledge reflects the level of not be correctly classification ,which is generated by the X under B on the upper and Lower Approximation, It describes the approximation degree to the set X of the knowledge division Internal and External[8], Especially when the property values have the strong correlation and no conflictdata, the proposed method, compared with information entropy methods, can get better decision tree, and reduces the computational volume.

Acknowledgements

This work was supported by Education Department of Anhui Province Key Natural Science Research Foundation of China under Grant No. KJ2010A352.

References

[1] Breiman L., Friedman J. H., Olshen R. A.. Classification and Regression Trees. Belmont: Wadsworth International Group;

1984.

[2] Yin D. S., Wang G. Y.,Wu Y. A.. Self-learning Algorithm for Decision Tree Pre-Pruning. Proceedings of the Third

International Conference on Machine Learning and Cybernetics, Shanghai, 26-29 August 2004.

[3] Miao D. Q., Wang J.. Based on rough set multivariate decision tree construction method. Software 1997; 8 (6): 425-431

[4] Shi Z. Z.. Knowledge Discovery. Beijing: Tsinghua University Press; 2002.

[5] Hu X. G., Zhang D. Y.. A new rough set based decision tree construction algorithm. Computer Science 2005; 32 (8.A) :7-9.

[6] Sha H. X..The research of several intelligent mining algorithms based on knowledge roughness and expand attribute

reduction.2004; 13-17 Fuzhou University, a master's degree thesis.

[7] Zeng H. L.. Rough set theory and its applications. Chongqing: Chongqing University Press; 1998.

[8] Pawlak Z. Rough Set Approach to Multi-Attribute Decision Analysis. European Journal of Operational Research 1994; 72(3):

443-459.