Professional Documents
Culture Documents
----------------------------------------------------------------------------------------------
1
The data for the CHD problem has been donated by
Desai Laboratories (Surat, Gujarat, India).
Incremental Learning aims to incorporate the effect of a d) Construct decision tree T2 using only
newly introduced (or discovered) class on an existing instances from S2.
learned system. However, to the best of our knowledge no e) Construct tree T3 from T1 and T2 which
methods for building attribute-incremental or class- retains all the learned concepts from T1,
incremental decision trees have been proposed so far. and adds whatever possible from T2 as
below:
1. If An+1 is the root node
2. Attribute Incremental Induction of Decision R2 of T2, then, examine
the decision nodes at
Trees each test value in R2. If
any of these decision
We have attempted to find a solution for building nodes N tests the same
attribute-incremental decision tree. Suppose we have a attribute as the root
decision tree for a problem, with which we are reasonably node R1 of T1, then set
satisfied. Now, suppose a new attribute comes to light, and the root node of T3 as
we would like to know the effect of this attribute on our R2, and replace the
problem. It may be that we do not have sufficient training subtree rooted at N by
instances with this new attribute to build a new decision tree the subtree rooted at R1
and the training instances are very slow in coming, so that in T3. Else, go to step
we do not want to wait a long time until we have collected e.2.
sufficient training data. Now, the new algorithm proposed, 2. Set T3 initially to be
called AIIDT (Attribute Incremental Induction of Decision same as T1. For each
Trees) imposes the effect of a newly discovered attribute branch in T2,
onto a previously learned decision tree. Moreover, in doing 3. If An+1 appears
so, we do not dispose any parts of the previously learned anywhere along the
decision tree. The proposed algorithm AIIDT just lets us branch, note the parent
learn whatever is possible without discarding any subtrees node P, the testvalue t at
from the older decision tree. This method will work whether P for which An+1 was
the new attribute is discrete or continuous. This method lets tested, and the child
us study the effect of the new attribute on our problem. node C of P along the
branch for testvalue t.
2.1 Algorithm AIIDT 4. For each occurrence of
P in T3:
Input: I. If t is a new
1. A Decision Tree T1 built using an instance set S1, testvalue for P,
that does not have values for the new attribute An+1. then add a new
2. A new instance set S2 that also has values for the branch from P
attribute An+1. with testvalue t
and copy the
Output: subtree rooted at
1. A new Decision Tree T3 that incorporates the effect An+1 (of T2)
of S2 onto already existing T1. below it.
Steps: II. Else, if P (of T3)
a) Let T1 be the older decision tree learned already has a
with attribute set A={A1, A2…An} using branch with
instances set S1. testvalue t
b) Let An+1 be the new attribute, whose effect leading to C,
we want to learn on our decision tree, i.e. then, introduce
we want to build a decision tree with decision node
attribute set A’={A1, A2…An, An+1}. An+1 as an
c) Let S2 be the new instance set that has additional test
instances with values for attribute An+1. between P and C.
----------------------------------------------------------------------------------------------
1
The data for the CHD problem has been donated by
Desai Laboratories (Surat, Gujarat, India).
In doing this, the
subtree at C is to Remark 1 : The algorithm naturally favors the concepts in
be retained, i.e. it T1 but this is simply because we too have more confidence
is not to be in the concepts of tree T1 which have been in existence and
replaced by the use since a long while, whereas the concepts in T2 are new
subtree of An+1 and subject to change as more and more instances are
rooted at C. collected for building tree T2.
T1 T2 T3
A1
A1
P
t A2
A3
A2 A3 TABLE 4
An+1
ATTRIBUTES USED IN THE CHD PROBLEM
P
TEST TYPE OF ATTRIBUTE
t
P t1
t2
1. serum cholesterol continuous
t
C
An+1 2. cholesterol/HDL ratio continuous
t1
t2 3. cholesterol/LDL ratio continuous
C 4. Diabetic discrete (boolean)
C 5. Male/Female discrete (boolean)
6. Age continuous
----------------------------------------------------------------------------------------------
1
The data for the CHD problem has been donated by
Desai Laboratories (Surat, Gujarat, India).
The analysis from these tests were considered satisfactory seventh attribute is the Homocystine attribute. The decision
most of the time. Afterwards, a new factor called tree thus obtained is shown in Figure 6.
Homocystine was also introduced. Its effect was also felt to
be significant. The doctors wanted to know how exactly the
Homocystine factor combined with the previous six well- Homo
-cystine
known and well-accepted factors in the determination of <16.07
coronary heart disease. >=16.07
serum High
cholester
The decision tree T1 obtained by top-down induction from <194.5 ol
CHD risk
High
LDL/HDL CHD risk
ratio
>=3.78
<3.78
serum
cholesterol
<213 Low CHD Diabetic
risk ?
>=213 no yes
chol/ HDL
ratio High Low CHD High
CHD risk risk CHD risk
<6
>=6
Figure 6 Decision Tree T2
LDL/ HDL
ratio High
CHD risk Finally, the third decision tree T3 obtained by applying
<4
>=4 algorithm AIIDT on T1 and T2 is shown in Fig. 2-6.
Low
Diabetic ?
CHD risk
no
yes
Male / High
Female ? CHD risk
female
male
High
CHD risk Age ?
<42.5 >=42.5
Low High
CHD risk CHD risk
----------------------------------------------------------------------------------------------
1
The data for the CHD problem has been donated by
Desai Laboratories (Surat, Gujarat, India).
References
Homo
-cystine
<16.07 >=16.0
7 [Mitchell, 1997] T. Mitchell. Machine Learning. McGraw
Hill, 1997.
serum High
cholest CHD
<213 erol risk [Fayyad and Irani] .U. M. Fayyad and K.B. Irani, On the
>=213 handling of continuous-valued attributes in decision tree
chol/
HDL High generation, Machine Learning 8, 87--102. 1992.
ratio CHD
risk
[Utgoff, 1989] P. E. Utgoff. Incremental induction of
>=6
<6 decision trees, 1989.
High
CHD
LDL/HD
L ratio risk [Utgoff, 1997] P. E. Utgoff. Decision tree induction based
>=4 on efficient tree restructuring, 1997.
<4
High
CHD
Age [Elomaa and Rousu, 1996] T. Elomaa and J. Rousu. Finding
risk <42.5 >=42.5
optimal multi-splits for numerical attributes in decision tree
Low High learning, 1996.
CHD CHD
risk risk
[Kohavi and Quinlan, 1999] R. Kohavi and R. Quinlan.
Figure 7 Decision Tree T3 obtained by AIIDT Decision tree discovery, 1999.
Continuous attributes were discretized using the method [Quinlan, 1986] J. R. Quinlan. Simplifying decision trees,
proposed by Fayyad and Irani for obtaining optimal binary 1986.
split[2].The labels on the two sides of the arrows in Figures [Webb, 2002] G. I. Webb. Furthur experimental evidence
5-7. show these values. The new decision tree thus obtained against the utility of Occam’s Razor. Journal of Artificial
using AIIDT has shown 97% accuracy over the CHD data Intelligence Research 4(397-417), 2002.
set.
[Amor et al., 2004] N. Amor, S. Benferhat and Z. Elouedi.
Naïve bayes vs decision trees in intrusion detection systems.
3 Conclusion ACM Symposium on Applied Computing, 2004.
Decision Trees are quick and effective means of learning [Esposito et al.,1992] F. Esposito, D. Malerba and G.
target concepts from given data. Algorithms like ID3 and Semeraro. A comparative analysis of methods for decision
C4.5 as well as the instance-incremental algorithms like tree generation, 1992.
ID5R can be used to solve a variety of problems for concept
learning and classification. The method proposed for binary
discretization [Fayyad and Irani, 1992 ] has worked out very
well for the CHD problem. The proposed new algorithm
AIIDT was successful in obtaining a good decision tree for
the attribute-incremental setting. To the best of our
knowledge, this is the first algorithm proposed for attribute-
incremental learning of decision trees.
----------------------------------------------------------------------------------------------
1
The data for the CHD problem has been donated by
Desai Laboratories (Surat, Gujarat, India).