Professional Documents
Culture Documents
A New Pruning Technique For The Fuzzy ARTMAP Neural Network and Its Application To Medical Decision Support
A New Pruning Technique For The Fuzzy ARTMAP Neural Network and Its Application To Medical Decision Support
net/publication/309427813
A New Pruning Technique for the Fuzzy ARTMAP Neural Network and Its
Application to Medical Decision Support
CITATIONS READS
0 20
3 authors:
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Shahrul Nizam Yaakob on 26 October 2016.
Abstract This paper describes a neural network-based classification tool that can
be deployed for data-based decision support tasks. In particular, the Fuzzy
ARTMAP (FAM) network is investigated, and a new pruning technique is pro-
posed. The pruning technique is implemented successively to eliminate those
rarely activated nodes in the category layer of FAM. Three data sets with differ-
ent characteristics are used to analyze its effectiveness. In addition, a benchmark
medical problem is used to evaluate its applicability as a decision support tool for
medical diagnosis. From the experiment, the pruning technique is able to improve
classification performances, as compared with those of to the original FAM net-
work, as well as other machine learning methods. More importantly, the pruning
technique yields more stable performances with fewer nodes, and results in a more
parsimonious FAM network for undertaking data classification and decision sup-
port tasks.
1. Introduction
The organizations of this paper are as follows: In Section 2, we discuss the dy-
namics of FAM and the problem of excess nodes in FAM. Section 3 explains the
data sets used, cross validation, and setting of FAM parameters, as well as the ex-
perimental results and discussion. Section 4 is about an analysis of the effective-
ness of FAM in medical diagnosis application. Section 5 gives the conclusions
and suggestions for further work.
I^wj
Tj = (1)
α + wj
M
I ≡ ∑I
i =1
i (3)
Recognition
F2 Layer
Map Field
F
ab
F1
Input Layer W jkab F a2 F b2
F0 Complement
Coding
ART ARTa
Input vector, a
Fig. 1 Fuzzy ART architecture (right) and Map Field architecture (left)
The next step is to find the maximum value of Tj, a competitive process, using
equation (4). Only one node in F2 with the highest value of Tj is selected. Reso-
nance is said to occur if the vigilance test (Equation (5) and (6)) is satisfied.
When the Jth category is chosen, yj = 1 and yj = 0 for j ≠ J. If Vj does not meet the
vigilance test, choice function Tj is set to zero for the duration of the input presen-
tation, and a new index j is chosen using Equation (4). All these operations occur
simultaneously in ARTa and ARTb.
Vj = I ^ wj / I (5)
5
Vj ≥ ρ (6)
The next stage is to find the value of X ab . If the J th of Fa2 node is active and
Kth of Fb2 is active, the value of Xab can be found using Equation (7). When both
ARTa and ARTb are active, and Xab ≠ 0, then the weight vector between Fa2 layer
and map field Fab is set according to Equation (8). If Xab = 0, then the vigilance
parameter of ARTa is increased such that the competitive process starts again, and
a new category that satisfies Eq. (5) is found. If there is no winner, a new node is
created and the value of input vectors is assigned as its weight. Learning ensues
by using Equation (9) whereby the weight vector wj of ARTa is updated to encode
the input pattern.
X ab = y b ^ w ab
j (7)
1 j = J;k = K
j =
w ab (8)
0 otherwise
( )
w (jnew) = β 1^ w (jold ) + (1 − β )w (jold ) (9)
In the test phase, Fuzzy ARTa receives an input vector. The category choice
and category match computations are the same as in the training phase. Therefore,
the output of map field Xab related to the Jth category of ARTa can be defined using
Equation (10). A link traced from the map filed to ARTb leads to the predicted tar-
get output.
ab ab
X =w (10)
j
In the pruning stage, every node j in the recognition layer is assigned with a pa-
rameter, Ωj. The pruning process only occurs during the training phase of FAM.
Each time when a new learning epoch begins, the value of Ωj is set to zero. Dur-
ing the learning epoch, Ωj changes to one when the associated node learns an input
pattern. This value remains as one irrespective of the number of times the node
becomes the winning node whereby its weight is updated. At the end of each
6
learning epoch, any node with Ωj equal to zero is pruned. These steps are repeated
for the next learning epoch until the learning phase is completed.
To analyze the effectiveness of the new pruning technique, three artificial data sets
are used. They are the Gaussian 2 Dimensional, Concentric, and Clouds data sets
from the ELENA project databases [10]. Fig. 4 and Fig. 5 are graphical
illustrations of Gaussian 2-D, Concentric, and Clouds, respectively. The Gaussian
2-D data set demonstrates a densely overlapped data distribution. The Concentric
data set has nested classes without overlapping. The Clouds data set shows
intersection of the class distribution, and has a high degree of nonlinearity in the
class boundaries. The details of the data sets are summarized in Table 1.
Gaussian 2D 1000 2 2
Concentric 1000 2 2
Clouds 1000 2 2
In this work, 10-folds cross validation is used where each data set is divided
into 10 mutually exclusive subsets, designated as G1, G2, G3,….G10, of equal size.
Then, FAM is trained and tested 10 times. The Percentage of Correct Classifica-
tion (PCC) is defined as the number of correct classification divided by the num-
ber of data available in the data set, as in Eq. (11). Note that σ(x,y)t=1 for correct
prediction, otherwise σ(x,y)t=0, and n refers to the number of data sample tested.
1 10
PCC
k
= 100 ∑ NCC
k
(11)
G
k =1
n
NCC = ∑ σ ( x, y )
t
(12)
k
t =1
The Standard Deviation Value (STD) is also calculated to reflect the distribution
of FAM results. Other FAM parameters are set as in Table 2.
7
Parameter Value
8
0.90
0.70
4
2
0.50
0
-8 -6 -4 -2 0 2 4 6 8
-2 0.30
-4
0.10
-6
5.50
4.50
3.50
2.50
1.50
0.50
-1.50
-2.50
-3.50
The FAM classification results, both the original and pruned versions, are shown
in Table 3. The pruning technique is able to improve accuracy rate by 5.1% to
6.3%, as compared with those of original FAM. For example, the classification
rate of Concentric increased by 6.3% to 97.6%. More importantly, the pruning
technique is able to reduce 34 to 136 nodes, as compared with those of original
FAM. This shows the effectiveness of the proposed pruning technique in produc-
ing a compact FAM network structure with improved classification performance.
The lower STD values also suggest the stability of the performance of the
pruned FAM network. In other words, the pruning technique makes the original
FAM less sensitive to the order of data presentation. This will make FAM more
suitable in on-line learning environment [4]. Note that the results of Gaussian 2-D
are low. This is caused by the densely overlapped region in data distribution. As
illustrated in Fig. 4, there is no clear delineation between the boundaries of the
two Gaussian 2-D classes. Nevertheless, the pruning technique is able to increase
the classification performance with fewer numbers of nodes.
In this work, the Wisconsin Breast Cancer (WBC) [3] data set is used to evaluate
the applicability of FAM with the proposed pruning technique as a medical deci-
sion support tool. The WBC data set contained 699 records of virtually assessed
nuclear features of fine needle aspirates from patients, with 458 benign and 241
9
malignant cases of breast cancer. The same training and test procedures as in Sec-
tion 3 were adopted to assess the FAM performance. Table 5 shows the classifica-
tion results of FAM (both original and pruned versions). It can be clearly seen
that while original FAM produced the lowest classification rate of 91.88%, the
pruned technique is able to improve its performance to 96.76%, which is the high-
est as compared with those of other machine learning methods (Table 5). The re-
sults demonstrate the effectiveness of the pruning technique in producing a parsi-
monious FAM network with improved classification performance.
5. Summary
In this paper, we have introduced a new pruning technique to reduce the number
of category nodes in the FAM network. The results obtained indicate that the
pruning technique can improve the classification performance result of FAM with
a more compact network structure. The results also are more stable (in terms of
STD) as compared with those from original FAM. Applicability of the proposed
approach as a decision support tool to medical diagnosis is also demonstrated us-
ing the WBC problem. The results, again, positively demonstrate that the pruned
FAM network, as compared with other machine learning methods, is able to pro-
duce high performance with a less complex network structure. The proposed prun-
ing technique can be incorporated with other FAM-based networks, e.g. Micro
ARTMAP and Bayesian ARTMAP. In addition, other pruning strategies can be
implemented and compared with the proposed technique. Besides, more bench-
mark and real data sets can be used to further ascertain the performance and stabil-
ity of the pruned network. All these constitute the direction of further work of this
research.
10
References
1. Abonyi, J. and F. Szeifert. Supervised fuzzy clustering for the identification of fuzzy classifi-
ers, Pattern Recognition Letter. Vol. 24, pp. 2195–2207.
2. Anagnostopoulos, G.C. et al. Reducing generalization error and category proliferation in ellip-
soid ARTMAP via tunable misclassification error tolerance: boosted ellipsoid ARTMAP. In-
ternational Joint Conference on Neural Networks, 2002, Vol.3, pp. 2650 – 2655.
3. Blake C.L. and Merz, C.J., UCI Repository of Machine Learning Databases,
http://www.ics.uci.edu/~mlearn/MLReposi-tory.html, University of California at Irvine. Cited
15 Mac 2008.
4. Bouchachia, Abdelhamid; Gabrys, Bogdan; Sahel, Zoheir, (2007) Overview of Some Incre-
mental Learning Algorithms, IEEE International Fuzzy Systems Conference, 2007, pp. 1 – 6.
5. Carpenter, G.A., Grossberg, S., Markuzon, N. and Reynolds, J.H.; Rosen, D.B. Fuzzy
ARTMAP: A neural network architecture for incremental supervised learning of analog mul-
tidimensional maps. IEEE Transactions on Neural Networks, Vol. 3, pp. 698 – 713.
6. Carpenter, G.A, Grossberg, S., Markuzon, N. and Reynolds, J.H.; Rosen, D.B. Fuzzy
ARTMAP: An adaptive resonance architecture for incremental learning of analog maps. In-
ternational Joint Conference on Neural Networks, 1992, Vol. 3, pp. 309 – 314.
7. Dong, C.S.J. and Loo, G.S.L. (2001). Flexible web-based decision support system generator
(FWDSSG) utilising software agents. 12th International Workshop on Database and Expert
Systems Applications, pp. 892 – 897.
8. Le, Q., Anagnostopoulos, G.C., Georgiopoulos, M. and Ports, K., An experimental compari-
son of semi-supervised ARTMAP architectures, GCS and GNG classifiers, IEEE Interna-
tional Joint Conference on Neural Networks, 2005. Vol.5. pp. 3121-3126,
9. Sanchez, E.G., Dimitriadis, Y.A., Cano-Izquierdo, J.M. and Lopez-Coronado, J. µARTMAP:
use of mutual inform-ation for category reduction in Fuzzy ARTMAP, IEEE Transactions on
Neural Networks, Vol.13, pp.58 – 69.
10. Verleysen, M., Bodt, E.D. and Wertz, V. UCL Neural Network Group,
http://www.dice.ucl.ac.be/neural-nets/Rese-arch/Projects/ELENA/elena.htm, Université ca-
tholique de L-ouvain. Cited 15 Mac 2008.
11. Vigdor B. and Lerner, B. The Bayesian ARTMAP, IEEE Transactions on Neural Networks,
Vol. 18, pp. 1628 – 1644.
12. Wee, C. Y., Paramesran, R., Takeda, F., Tsuzuki, T., Kadota, H., Shimanouchi, S. Classifica-
tion of rice grains using Fuzzy Artmap neural network, Asia-Pacific Conference on Circuits
and Systems, Vol. 2, pp. 223-226.
13. Zhong M., Rosander, B., Georgiopoulos, M., Anagno-stopoulos, G.C., Mollaghasemi, M.
and Richie, S., Experiments with Safe ARTMAP and Comparisons to Other ART Networks,
International Joint Conference on Neural Networks, 2006. pp. 720-727.