Professional Documents
Culture Documents
1871
1.0 I I
tively. An interval is represented by its lower limit
and upper limit as A = [aL,U’] where uL and U’
are the lower limit and upper limit of the interval
A, respectively. Therefore the interval vector X , is
represented as
X p (Xpl,...,Xpn)
L
= ([z,”,,zpUlIi-..,[zpn,zPU,I). (2)
Let us define the input-output relation of a three-
layer feedforward neural network with n input units,
n H hidden units and c output units when the inter-
val input vector X, is presented. By extending the
standard back-propagation network to the case of the
interval input vector X , , the input-output relation
of each unit is defined by interval arithmetic[6] as Fig.1. Activation function f(.)
follows:
Input unit:
o,, = [o,”,, 0 3 = x,i = [z”.
pa1 ‘
.pi],
O ( X ) = ( 0 1 ( X ) ., . . , O , ( X ) ) . Because O ( X )is cal-
culated by (3)-(9) based on interval arithmetic (i.e.,
i = 1,2,...,n, (3) the input-output relation of each unit satisfies the
inclusion monotonicity of intervals[6]), the following
proposition holds.
[Proposition 11
For any pair of n-dimensional interval vectors
X and Y such that X E Y (i.e., Xi 2
n n E, i = 1 , 2 , . . . , n), the following inclusion relation
holds.
o k ( x ) Ok(Y), k = 1 , 2 , ’ , c.
” (10)
1872
We define a cost function for the p t h training pat- IV. Application to Medical Diagnosis
tern ( X , , t , ) as A. Data of Hepatac Daseases
C C
We applied the proposed approach to the medical di-
agnosis of hepatic diseases in Tanaka et aZ.[5]. The
k=l k=l classification problem of hepatic diseases has five
classes (Class 1: Healthy person, Class 2: Hepatoma,
where Opk= [o$, OFk] is the interval output from Class 3: Acute hepatitis, Class 4: Chronic hepati-
the k-th output unit calculated by (3)-(9). tis, Class 5: Liver cirrhosis), 568 persons and 20 at-
In the same manner as the back-propagation algo- tributes (i.e., 20 medical inspections of each person).
rithm[4], the weights W k j and wjj are changed by the We applied the proposed method to these data on the
following rules: same conditions as Tanaka et al.[5]. That is, we used
AWkj(t + 1 ) = q(-dep/dWkj) + a A w k j ( t ) , (14)
the following conditions in computer simulations:
( 1 ) The given data were divided into two parts:
A w j i ( t + 1 ) = q(-de,/dwji) + a A w j i ( t ) ,
(15) Data of 468 persons were used for training neural
networks, and those of 100 persons were used for
where the derivatives de,/dwkj and de,/dwji are testing the trained neural networks.
calculated as follows: ( 2 ) Seven attributes in Table 1, which had been
selected in [5], were used in the learning of neural
networks. Each attribute value was pre-processed
and transformed into one of the pre-specified discrete
values in the unit interval [0,1] as shown in Table
1. For example, if the attribute value of SP is in
the interval [5.6,6.5], it is converted into 0.25. The
number of missing values are also shown in Table 1.
From Table 1, we can see that a number of attribute
values are missing.
B. Classificataon Rules
In Section 3, we have already proposed a learning al-
gorithm of a three-layer feedforward neural network.
Because the output vector from the trained neural
network is an interval vector, we should specify clas-
sification rules based on the interval output vector.
In computer simulations, we employed the following
heuristic classification procedure.
[Classification of the input vector X,]
Step 1: Calculate the interval output vector 0,
from the trained neural network corresponding
to the interval input vector X , .
Step 2: Calculate the index set Q as follows:
C
4 = {kloFk 2 0.5, k = 1 , 2 , ” ‘ , c } . (22)
Step 3: ( 1 ) If 4 has a single element, say k*, then
classify X, as Class k*.
(2) If 9 has more than one element, then
reject the classification of X,.
e If wji < 0, (3) If \E has no element, then classify X,
C as Class h such that
orh = m a x { o ~ k l l c = 1 , 2 , . . . , c }. (23)
C. Simulation Results
We trained a neural network with 7 input units, 30
hidden units and 5 output units. In computer sim-
C
ulations, initial values of the weights and the biases
were randomly specified in [-1,1], and the learning
constant 11 and the momentum constant a were spec-
ified as q= 0.075 and a = 0.9. Ten trials were per-
formed with different initial weights and biases. The
following average performance of the trained neural
1873
network after 800 epochs was obtained. V. Conclusion
Classification rate: 62.l[%] We proposed a neural-network-based approach to the
Error rate: 15.2[%] multi-class classification problem of incomplete data
Reject rate: 22.7[%] with missing values. In the proposed approach, miss-
ing attribute values were represented by intervals.
From the comparison of these results with the fol- Therefore our problem was reduced to the multi-class
lowing results by the rule-based fuzzy classification classification problem of interval vectors. We pro-
system in Tanaka et a1.[5], we can see that the error posed an architecture and a learning algorithm of
rate was significantly decreased from 23% to 15.2% neural networks for that problem. The proposed ap-
by the proposed method. proach was applyed to the medical diagnosis of hep-
Classification rate: 62[%] atic disease in Tanaka et aZ.[5], and its performance
E h o r rate: 23[%] was compared with that of the rule-based fuzzy clas-
Reject rate: 15[%] sification system.
In order to examine the relation between the per-
formance of the trained neural network and the num-
ber of iterations of the learning algorithm, we mon- References
itored the average error rate in the learning of the [l] S.Ahmad and V.Tresp: “Classification with Miss-
neural network every 40 epochs. The simulation re- ing and Uncertain Inputs”, Proc.IEEE-ICNN’93,
sults are summarized in Fig.2. From this figure, we pp.1949-1954 (1993).
can see that the performance for the test data was
almost the same after 200 epochs whereas that for [2] H.Ishibuchi and H.Tanaka: “An Extension of The
the training data was gradually improved. BP-Algorithm to Interval Input Vectors”, Proc. of
IJCNN’91-Singapore, pp.1594-1599 (1991).
-
I1 0 2
0- 0.00
- 0.33
-
0.67
100 101 150 151 200 201 250 251 500 501 00 -
1.00
-
-
ChE 3 3
GPR -0.00
0 25 26
0.20
- 0.40
- 0.60
-
100 101 200 201 500 501 1000 1001 00
0.80
-1.00
0 1
Lympho
0.00 0.20
-
0.40
-
0.60
0 N 20.0 20.1 40.0 40.1 60.0 60.1 00 -0.80 1.00
6 3
Al-%
-
0.00
-
0.33
-
0 2.5 2.6 3.7 3.8 5.0 5.1 CO
0.67
- 1.00
27 5
AFP
-
0 20 21
0.00
-
0.00
0.25
-
0.33
100 101 200 201
0.50
-
0.67
0.75
1.00
-
1000 1001 cm
1.00
95 45
1874