You are on page 1of 8

COUNTERPROPAGATION NEURAL NETWORKS IN

STRUCTURAL ENGINEERING

By Hoijat Adeli, I Member, ASCE, and Hyo Seon Park2

ABSTRACT: Neural network computing has recently been applied to structural engineering problems. Most
of the published research is based on a back-propagation neural network (BPN), primarily due to its simplici.ty.
The back-propagation algorithm, however, has a slow rate of learning and is the~efo~e impractical for learn~ng
of complicated problems requiring large networks. In this paper, we present applIcatIOn of c<;>unterpr?pagatlon
neural network (CPN) with competition and interpolation layers in structural analysis and desIgn. To cIrcumvent
the arbitrary trial-and-error selection of the learning coefficients encountered in the counterpropagation a!-
Downloaded from ascelibrary.org by Technische Universitat Munchen on 07/04/15. Copyright ASCE. For personal use only; all rights reserved.

gorithm, a simple formula is proposed as a function of the iteration number and excellent conve~gence IS
reported. The CPN is compared with the BPN using two structural. engineering. examples reported I~ rece~t
literature. We found superior convergence property and a substantIal decrease In the central processIng u.nIt
(CPU) time for the CPN. In addition, CP~ was ~pplied to two new examples In the area of steel des~gn
requiring large networks with thousands of lInks. It IS shown that CPN can learn complIcated structural desIgn
problems within a reasonable CPU time.

INTRODUCTION determined such that the domain error is reduced in the sub-
sequent iterations. The application of the new algorithm to
In the first structural engineering application of neural net- structural design problems demonstrates its superior conver-
works published in an archival journal, Adeli and Yeh (1989) gence property.
presented a model of machine learning in engineering design Gunaratnam and Gero (1994) discuss the effect of repre-
based on the concept of self-adjustment of internal control sentation on the performance of neural networks in structural
parameters and perceptron. The problem of structural design engineering applications using the error back-propagation al-
was cast in a form to be described by a perceptron without gorithm. They suggest that dimensional analysis provides a
hidden layers. suitable representation framework for training the input/out-
Vanluchene and Sun (1990) discuss the use of back-prop- put pattern pairs.
agation learning algorithm (Rumelhart et al. 1986) in struc- Back-propagation seems to be the most utilized neural net-
tural engineering. Hung and Adeli (1991a) present a two- work algorithm in civil engineering disciplines. This is pri-
layer neural network learning model for engineering design marily due to its simplicity. A detailed discussion of the back-
by combining the perceptron with a single-layer AND neural propagation neural network (BPN) algorithm with a view on
net. They reported improvement in the rate of learning com- engineering applications can be found in Hegazy et al. (1994).
pared with the single-layer perceptron learning model. Sev- They also provide a structured framework for developing
eral other researchers have applied neural networks, mostly practical neural network applications.
back-propagation algorithms, in structural engineering and The back-propagation algorithm, however, has a slow rate
mechanics and related engineering problems (Hajela and Berke of learning. Consequently, it cannot be readily applied to
1991; Ghaboussi et al. 1991; Masri et al. 1993; Kang and large problems. One approach to improve the learning per-
Yoon 1994; Messner et al. 1994; Stephen and Vanluchene formance is development of parallel algorithms on multipro-
1994; Elkordy et al. 1994; Hurson et al. 1994). cessor machines (Adeli 1992a,b). Hung and Adeli (1993) pre-
Hung and Adeli (1994a) present a multilayer neural net- sent parallel back-propagation learning algorithms employing
work development environment for effective implementation the microtasking and vectorization capabilities of vector MIMD
of learning algorithms for the domain of engineering design machines such as the Cray YMP8/864 supercomputer (Saleh
using the object-oriented programming paradigm (Adeli and and Adeli 1994).
Yu 1993; Yu and Adeli 1993). It consists of five primary How to improve the convergence speed of neural network
components: learning domain, neural nets, library of learning
learning algorithms is of significant importance and is cur-
strategies, learnng process, and analysis process. These com-
rently being actively researched (Hung and Adeli 1991b; Adeli
ponents have been implemented as five classes in the object-
and Hung 1993a,b; Hung and Adeli 1994b). Most recently,
oriented programming language C++. The library of learn-
ing strategies includes the generalized delta rule with error Adeli and Hung (1994) presented an adaptive conjugate gra-
back-propagation. dient learning algorithm for training of multilayer feedfor-
Adeli and Zhang (1993) present an improved perceptron ward neural networks. The problem of arbitrary trial-and-
learning algorithm by introducing an adjustment factor in error selection of the learning and momentum ratios en-
each self-modification iteration of the original perceptron countered in the momentum back-propagation algorithm is
learning model. The adjustment factor in each iteration is circumvented in the new adaptive algorithm. Instead of con-
stant learning and momentum ratios, the step length in the
inexact line search is adapted during the learning process
IProf., Dept. of Civ. Engrg., Ohio Slate Univ., 470 Hitchcock Hall,
2070 Neil Ave., Columbus, OH 43210. through a mathematical approach. Thus, the new adaptive
°Ph.D. Candidate, Dept. of Civ. Engrg., Ohio State Univ., 470 Hitch- algorithm provides a more solid mathematical foundation for
cock Hall, 2070 Neil Ave., Columbus, OH. neural network learning. The algorithm has been applied to
Note. Associate Editor: David Darwin. Discussion open until January two different domains: engineering design and image rec-
I, 1996. To extend the closing date one month, a written request must ognition. It is shown that the adaptive neural networks al-
be filed with the ASCE Manager of Journals. The manuscript for this
paper was submitted for review and possible publication on August 13,
gorithm has superior convergence property compared with
1993. This paper is part of the Journal of Structural Engineering, Vol. the momentum back-propagation algorithm.
121, No.8, August, 1995. ©ASCE, ISSN 0733-9445/95/0008-1205-1212/ In this paper, we explore the use of counterpropagation
$2.00 + $.25 per page. Paper No. 6765. neural networks (CPN) in structural engineering. In contrast
JOURNAL OF STRUCTURAL ENGINEERING / AUGUST 1995/1205

J. Struct. Eng. 1995.121:1205-1212.


to back-propagation, which is a supervised learning algo- stances X and desired output Yare presented to the network
rithm, counterpropagation network is a combination of su- at the input and interpolation layers, respectively. Thus, the
pervised and unsupervised (self-organizing) mapping neural numbers of nodes in the input and interpolation layers cor-
networks (Hecht-Neilsen 1987a,b). The performances of back- respond to the numbers of elements in the vectors X and Y,
propagation and counterpropagation learning algorithms are respectively. The vector of computed output is represented
compared for structural engineering applications. Four ex- by Y ' . A counterpropagation network is trained in two suc-
amples are presented. The first two examples are taken from cessive steps. The counterpropagation learning algorithm is
the literature for the sake of comparison. They are the con- presented schematically in Fig. 2.
crete beam design and the prediction of the maximum bend- The first step is performed between the input layer and the
ing moment in a simply supported plate (Vanluchene and Sun competition layer. For each pair of training instance and de-
1990; Gunaratnam and Gero 1994). The other two examples sired output (X, V), each component of training instance, Xj'
are new. The first one is the prediction of the critical elastic is presented to the corresponding node of the input layer. Let
lateral torsional buckling moment of beams. The second one V j be the arbitrary initial weight vector assigned to the links
Downloaded from ascelibrary.org by Technische Universitat Munchen on 07/04/15. Copyright ASCE. For personal use only; all rights reserved.

is the prediction of the moment-gradient coefficient Cb for between input nodes and the jth node in the competition
doubly and singly symmetric steel beams. layer. The transfer function for the competition layer is de-
fined by the Euclidean distance between the weight vector V j
COUNTERPROPAGATION and the training instance X as follows:
The back-propagation algorithm has been covered in lit- dj = IIDj - XII (1)
erature extensively (Rumelhart et al. 1986; Hung and Adeli
1993 a,b) and therefore will not be described here. We will, where IIXII = VljlxY. For the given training instance X,
however, review the counterpropagation algorithm briefly in each node in the competition layer competes with other nodes
this section (Woods 1988). The topology of a counterprop- and the node with the shortest Euclidean distance to X wins.
agation network consists of three primary layers: input layer, As a result of the competition, the output of the winning node
competition layer, and interpolation layer (Fig. 1). is set to 1.0 and outputs of the other nodes are set to O. Thus
For training the network, the vectors of input training in- the output of the jth node in the competition layer, Zj' is
given by
INPUT (X) COMPETITION INTERPOLATION INPUT (Y) OUTPUT (Y')
LAYER LAYER
Z = {1.0 if d j < d for all i
j (2)
I 0 otherwise
A weight Uji assigned to the link connecting the node j in
the competition layer and the node i in the input layer is
adjusted according to the Kohonen (1988) learning rule
Ujj(n + 1) = UJj(n) + a[Xj - Ujj(n)]Zj (3)
where n = iteration number; and a = learning coefficient.
For the learning coefficient, Hecht-Neilsen (1988) suggests a
number in the range of 0 < a ~ 0.8. In this paper, however,
we define the learning coefficient, a, as a function of iteration
number in the following form
a = lI(n + 1)2 (4)
After the weight vector V j of the competition layer stabi-
lizes, the interpolation layer starts to learn the desired output.
The weight assigned to the link between the winning node i
FIG. 1. Topology of Counterpropagation Network
in the competition layer and the jth node in the interpolation
layer, Vji, is adjusted according to the learning rule suggested
Training Phase (step 1)
by Grossberg (1982)
Verification of Untrained
instances (step 2)
Vjj(n + 1) = Vjj(n) + b[Vjj(n) + Yj]Zj (5)
where b = learning coefficient. Hecht-Neilsen (1988) sug-
gests a number in the range of 0 < a ~ 1.0. But, again in
this work we use (4), which we find more reasonable based
on numerical experimentations. The interpolation layer uses
a weighted summation function as a transfer function. The
jth element of the computed output of the network, Yj, is
determined by
(6)

During training (step 1 in Fig. 2) only one node of the


competition layer can win, and the corresponding output is
set to 1.0. After connection weights in a network stabilize,
the performance of the network can be tested by using un-
trained instances. During the verification of untrained in-
stances (step 2 in Fig. 2) the number of winning nodes in the
competition layer can be more than one. The nonzero outputs
FIG. 2. Counterpropagation Learning Algorithm of winning nodes are set such that the node associated with
12061 JOURNAL OF STRUCTURAL ENGINEERING 1 AUGUST 1995

J. Struct. Eng. 1995.121:1205-1212.


TABLE 1. Learning Results for Concrete Beam Design Problem
Gunaratnam and Gero (1994)
BPN (1-3-1) BPN (1-8-1) CPN (1-21-1)
Untrained instance ~ Computed Error Computed Error Computed Error
number f~bd2 p(fylf~) output (%) output (%) output (%)
(1 ) (2) (3) (4) (5) (6) (7) (8) (9)
1 0.1761 0.2267 0.2272 0.22 0.2259 0.35 0.2239 1.25
2 0.1673 0.2133 0.2147 0.65 0.2128 0.23 0.2444 0.53
3 0.1596 0.2000 0.2037 1.82 0.2013 0.65 0.2002 0.09
4 0.1894 0.2400 0.2460 2.44 0.2452 2.12 0.2448 1.97
5 0.1660 0.2100 0.2129 1.36 0.2109 0.43 0.2080 0.97
6 0.1770 0.2280 0.2285 0.22 0.2273 0.31 0.2250 1.33
7 0.0922 0.1100 0.1056 4.17 0.1076 2.23 0.1114 1.26
Downloaded from ascelibrary.org by Technische Universitat Munchen on 07/04/15. Copyright ASCE. For personal use only; all rights reserved.

8 0.1332 0.1650 0.1656 0.36 0.1619 1.91 0.1596 3.38


9 0.0910 0.1080 0.1039 3.95 0.1062 1.69 0.1098 1.68
10 0.1445 0.1800 0.1820 1.1 0.1786 0.78 0.1732 3.93
CPU time on Cray
YMP8/864 - - 23.50 sec - 26.65 sec - 0.008 sec -

TABLE 2. Learning Results for Simply Supported Plate


Input Bending about X-axis Bending about Y-axis
COMPUTED OUTPUT COMPUTED OUTPUT
Gunaratnam and Gunaratnam and
Plate Load Gero (1994) Gero (1994)
Dimension Position Desired Output BPN (4-6-6-6) CPN (4-30-6) Desired Output BPN (4-6-6-6) CPN (4-30-6)
X Y X Y Mx X Y Mx X Y Mx X Y My X Y My X Y My X Y
1.00 1.00 0.75 0.50 0.30 0.73 0.50 0.32 0.74 0.52 0.31 0.63 0.44 0.29 0.73 0.50 0.30 0.74 0.52 0.31 0.63 0.44
1.00 1.00 0.90 0.80 0.21 0.88 0.77 0.17 0.91 0.83 0.20 0.78 0.76 0.19 0.88 0.77 0.12 0.91 0.83 1l.21 0.78 0.76
0.90 1.00 0.18 0.20 0.25 0.20 0.22 0.26 0.18 0.22 0.25 0.20 0.22 0.25 0.20 0.22 0.25 0.18 0.22 0.25 0.20 0.22
0.90 1.00 0.22 0.80 0.26 0.25 0.73 0.27 0.24 0.77 0.27 0.21 0.74 0.26 0.25 0.73 0.26 0.24 0.77 0.26 lUI 0.74
1.00 0.80 0.60 0.12 0.22 0.57 0.14 0.22 0.58 0.14 0.22 0.57 0.14 0.26 0.57 0.14 0.26 0.58 0.14 0.26 0.57 0.14
1.00 0.80 0.80 0.20 0.26 0.77 0.22 0.25 0.76 0.22 0.25 0.77 0.23 0.27 0.77 0.22 0.27 0.76 0.22 0.27 0.77 0.23
0.70 1.00 0.35 0.50 0.33 0.35 0.50 0.32 0.35 0.49 0.32 0.33 0.50 0.30 0.35 0.50 0.30 0.35 0.50 0.29 0.33 0.50
0.70 1.00 0.56 0.90 0.21 0.54 0.88 0.25 0.46 0.74 0.22 0.46 0.62 0.21 0.54 0.88 0.27 0.46 0.74 0.23 0.46 0.62
0.60 1.00 0.30 0.75 0.30 0.30 0.73 0.30 0.30 0.74 0.29 0.30 0.74 0.28 0.30 0.73 0.28 0.30 0.74 0.28 1J.30 0.74
1.00 0.60 0.90 0.48 0.20 0.88 0.47 0.25 0.85 0.41 0.25 0.81 0.51 0.21 0.88 0.47 0.26 0.85 0.41 0.25 0.81 0.51
0.50 1.00 0.10 0.20 0.25 O.ll 0.22 0.25 0.07 0.23 0.26 0.19 0.22 0.21 0.11 0.22 0.22 0.07 0.23 0.23 0.19 0.22
1.00 0.50 0.20 0.38 0.23 0.22 0.36 0.24 0.18 0.32 0.24 0.34 0.19 0.26 0.22 0.36 0.27 0.18 0.32 0.28 0.34 0.19

3Or----------------,
l
~ Competition layer
en 0----<>

.-~ Interpolation layer


~20 - M -CPN
ig- .-_4 M;-BPN IG&G,l994)
- M -CPN
.. --~ M;- BPN IG &G, 1994)

110
~
E

! o 0'-~--=~---20""""""-"""3"'0"""-""""'40
No. of heratlon
4 5 6 7 8 9 10 11 12
FIG. 3. Convergence Behavior of Counterpropagation Untrained instance number

FIG. 4. Comparison of Learning Results for BPN and CPN


the weight vector closest to the given untrained instance has
the largest output. However, the sum of the outputs of win-
ning nodes in the competition layer remains equal to 1.0. By where d k = Euclidean distance between the weight vector
letting n win be the number of winning nodes in the competition assigned to the kth winning node and the untrained instance.
layer and Sw be the set of winning nodes, we define Zj as Thus, the output of the network for untrained instances is
follows: calculated by (6) with more than one nonzero output from
the competition layer, Zj.
The selection of the number of the winning nodes is prob-
lem dependent and plays an important role in the perfor-
(7) mance of a counterpropagation network. Different numbers
of winning nodes in the competition layer (different numbers
of nonzero outputs from the competition layer) show different
performance.
JOURNAL OF STRUCTURAL ENGINEERING 1 AUGUST 1995/1207

J. Struct. Eng. 1995.121:1205-1212.


Mal 1; ~Mal Length

Depth

~t1 i'~ Unit weight


I I
(a) L

FIG. 5. Counterpropagation for Lateral Torsional Buckling Example

TABLE 3. Training Data for Lateral Torsional Buckling of Steel 2OO00r--~-----r------r---------,


Downloaded from ascelibrary.org by Technische Universitat Munchen on 07/04/15. Copyright ASCE. For personal use only; all rights reserved.

(a)
Beams
Training Training Training - Computed output for simple beam from CPN
•.........• Desired output for simple beam from theory
instance Designation instance Designation instance Designation
(1) (2) (1 ) (2) (1) (2)
1 W36x300 11 W 33 x 141 21 W 27 x 114
2 W36 x 260 12 W 33 x 118 22 W 27 x 102
3 W 36 x 230 13 W 30 x 211 23 W 27 x 84
4 W 36 x 194 14 W 30 x 173 24 W 24 x 131
5 W 36 x 170 15 W 30 x 132 25 W24 x 104
6 W 36 x 150 16 W 30 x 124 26 W24 x 84
7 W 36 x 135 17 W 30 x 108 27 W 24 x 69
8 W 33 x 241 18 W30x90 28 W 24 x 55 ~
e or--~-_ .....
_-_.::.....-'----:.::::..~.....::::::t
9 W 33 x 201 19 W 27 x 178
10 W 33 x 152 20 W 27 x 146 CD 20.00
§ 10.00
E
~ 0.000l-.--'--'-10---20----3·0---40---50--....
60
TABLE 4. Verification Data for Lateral Torsional Buckling of Steel
Beams Untrained instance number

Verifica- Verifica- Verifica-


FIG. 6(a). Learning Results for Lateral Torsional Buckling of Sim-
ple Beams by CPN
tion tion tion
instance Designation instance Designation instance Designation
50000
(1) (2) (1 ) (2) (1) (2) (b)

1 W36 x 280 5 W 30 x 191 9 W 24 x 117 - Computed output for fixed beam from CPN
2 W 36 x 160 6 W 30 x 116 10 W 24 x 76 E 40000 - - Desired output for fixed-beam from theory
3 W 33 x 221 7 W 27 x 161 :Z
~
4 W 33 x 130 8 W 27 x 94 1:: 30000
CD
E
0
E
If the number of winning nodes is set to one, the winning lD
20000
node is the node whose weights are the closest to the input !:i
vector in a Euclidean sense. In this case, the CPN works as ....•<:B 10000
()
a simple nearest-neighbor classifier.
By setting the number of winning nodes to a number greater
than one, however, the accuracy of the mapping approxi-
l 0
mation can be improved significantly. In this case, the CPN
gCD 20.00
works as an interpolator with multiple winning nodes. The E 10.00
:::l
E
number of winning nodes in the competition layer is selected .~ 0.00
based on the set of input vectors (training instances) in any ::< 0 10 20 30 40 50 60
Untrained instance number
given problem. The optimum number of winning nodes can
be obtained from experience and through numerical experi- FIG. 6(b). Learning Results for Lateral Torsional Buckling of Fixed-
mentation only. In this research, for each problem we find Beams by CPN
the number of winning nodes that minimizes the maximum
error between the computed and desired outputs using a set
of verification instances. Concrete Beam Design Example
This example is to find a mapping neural network that
COMPARISON OF BACK-PROPAGATION
defines the relationship between M" (ultimate bending mo-
AND COUNTERPROPAGATION ment) and d (depth of a reinforced concrete beam with rec-
To compare the performances of back-propagation and tangular cross section). The relationship requires six varia-
counterpropagation neural network algorithms, two examples bles: p (reinforcement ratio), .f. (yield stress of reinforcing
solved by Vanluchene and Sun (1990) are considered. A three- steel), f~ (compressive strength of concrete). b (width of the
layer back-propagation network was used in these examples. rectangular beam), M,,, and d. This problem can be expressed
Subsequently, Gunaratnam and Gero (1994) reported im- in terms of two nondimensional variables pf"!f;. and M,)
proved performance for these examples by using dimension- f~bd2 as follows: .
less parameters. Hence, we use the dimensionless parameters
in both examples for comparisons of learning results and cen- (~)
tral processing unit (CPU) time.
12081 JOURNAL OF STRUCTURAL ENGINEERING 1 AUGUST 1995

J. Struct. Eng. 1995.121:1205-1212.


The topology of the counterpropagation network used in
this work consists of one input node, 21 competition nodes,
and one interpolation node, identified as CPN (1-21-1) in
Table 1. It took only 0.008 CPU seconds for both training of
the same 21 instances and testing the 10 untrained instances
on Cray YMP8/864. Thus, for this example the counter-
propagation algorithm is nearly 3,000 times more efficient
than the back-propagation algorithm, where two networks
produce comparable errors in testing of untrained instances.
Fig. 3 shows the convergence behavior for both competition
and interpolation layers.
We observe that the proposed equation for estimating the
learning coefficients a and b [(4)] provides rapid convergence.
Downloaded from ascelibrary.org by Technische Universitat Munchen on 07/04/15. Copyright ASCE. For personal use only; all rights reserved.

Furthermore, the problem of arbitrary trial-and-error selec-


tion of learning and momentum ratios encountered in the
back-propagation learning algorithm is circumvented since
the coefficients a and b are computed automatically in each
iteration without any user input.

Simply Supported Plate Example

llITT As an application of neural networks to structural analysis


problem, this example is the prediction of the locations and
magnitudes of maximum moments in a simply supported rec-
FIG. 7. Counterpropagation for Moment-Gradient Coefficient Ex-
tangular plate subjected to a unit concentrated load some-
ample where on the plate. In this example, each instance consists
of four components: the dimensions of the plate in the X-
and Y-directions and the locations of the load in the corre-
where 'P = function defining the relationship. The objective sponding coordinates. The desired output vector has six com-
of neural networks is to find the relationship function 'P. We ponents: the maximum bending moments M., and M" and the
first used the same network topology used by Gunaratnam X and Y coordinates of their locations. The back-propagation
and Gero (1994): one input node, one output node, and a results shown in Table 2 are from Gunaratnam and Gero
hidden layer with three nodes, identified as BPN (1-3-1) in (1994) for a network with four input nodes, two six-node
Table 1. Twenty-one training instances and 10 untrained in- hidden layers, and six output nodes. Thirty training instances
stances given in Table 1 were used for training and verification and 12 untrained instances from finite-element analysis were
of the network. used for training and verification of the network, respectively.
To reduce the error, we also used a network with one input In our counterpropagation network, we used four input
node, one output node, and one hidden layer having eight nodes, one competition layer with 30 nodes, and six inter-
nodes, identified as BPN (1-8-1) in Table 1. Training back- polation (output) nodes and the same training and verification
propagation network with instantaneous updating took 23.50 instances. During the verification step, according to (7) two
and 26.65 CPU seconds on a Cray YMP8/864, for the two winning nodes in the competition layer with the best perfor-
aforementioned networks, respectively. The percentage er- mance were selected for calculating the output of untrained
rors for the verification instances were reduced when the three instances. The learning results are compared with the results
hidden nodes were replaced with eight hidden nodes. from the back-propagation network in Table 2. Fig. 4 indi-

TABLE 5. Learning Results for Moment-Gradient Coefficient Corresponding to Eq. (13)


Input Pattern Input Pattern
Computed Desired Error Computed Desired Error
K p I/Ix 13 output output (%) K p I/Ix 13 output output (%)
0,62 0,93 0.06 ~I I I 0.00 1.34 0.41 fJ.OII -I I I 0.00
0.62 0.93 0.06 -0.8 1.11 1.11 0.01 1.34 0.41 0,011 -0.11 1.11 1.11 0.00
0.62 0.93 0.06 -0.6 1.243 1.242 0,03 1.34 0.41 0,08 -0.6 1.245 1.245 0.00
0,62 0,93 0.06 -0.4 1.404 1.403 0.08 1.34 0.41 0.08 -0.4 1.41 1.411 (UJI
0.62 0.93 0.06 -0.2 1.599 1.596 0.20 1.34 0.41 fJ.08 -0.2 1.617 1.617 fUl4
0.62 0.93 0.06 0 1.823 1.816 0.39 1.34 0.41 0.08 0 1.873 1,1174 0.09
0.62 0.93 0.06 0.2 2.031 2.022 0.43 1.34 0.41 0.08 0.2 2.182 2.1117 0.22
0,62 0.93 0,06 0.4 2.004 2.066 2.99 1.34 0.41 0.011 0.4 2.525 2.539 0.57
0.62 0.93 0.06 0.6 1.64 1.765 7.011 1.34 0.41 fJ.OII 0.6 2.833 2.873 1.40
0.62 0.93 0.06 0.11 1.311 1.357 3.41 1.34 0.41 (J.OII 0.11 2.9116 3.074 2.88
0.62 0.93 0.06 I 1.061 1.066 0.45 1.34 0.41 0.08 I 2.918 3.053 4.40
0.74 0.71 0.06 -I I I 0.00 1.R7 0.77 0.07 -I I I 0.00
0.74 0.71 0.06 -0.8 1.11 1.11 0.00 1.117 o.n 0.07 -0.11 1.11 1.11 0,00
0.74 0.71 0.06 -0.6 1.244 1.244 0.00 1.117 o.n 0.07 -0,6 1.242 1.242 fJ.OI
0.74 0.71 0.06 -0.4 1.4011 1.4011 0.01 1.117 0.77 0.07 -0.4 1.403 1.403 fJ.OI
0.74 0.71 0.06 -0,2 1.61 1.61 0.03 1.117 0,77 0.07 -0.2 1.597 1.597 0.03
0.74 0.71 0.06 0 1.1156 1.855 0.06 1.117 0.77 0.07 0 1.R21 1.1122 (J.OII
0.74 0.71 0,06 0,2 2.136 2.134 0.12 1.87 o,n 0.07 0.2 2.0411 2.052 0.20
0.74 0.71 0.06 0.4 2.4 2.395 0.21 1.117 0.77 0.07 0.4 2.187 2.196 0.41
0.74 0.71 0,06 0.6 2.504 2.499 0.23 1.87 0.77 0,07 0.6 2.093 2.096 0.15
0.74 0.71 0.06 0.8 2,323 2.309 0,61 1.117 0.77 0.07 0.11 1.786 1.753 1.115
0.74 0.71 0.06 I 1.996 1.962 1.73 1.117 0.77 0.07 I 1.4511 1.3911 4.27

JOURNAL OF STRUCTURAL ENGINEERING 1 AUGUST 1995/1209

J. Struct. Eng. 1995.121:1205-1212.


~ 3.5 r--....--,..-----,.-----,---r-......- ......- ......- ......---, American Institute of Steel Construction (AISC) Allowable
.[
(I)
~Eq.12 Stress Design (ASD) specifications (AISC 1989) are used as
~ 3.0 -- set-1 training instances (Table 3). The input for training and ver-
.._..__.. set-2

!,g 2.5
---
---
set-3
set-4
./ ification instances (Table 4) are the W shape designations
(the nominal depth and weight per unit length), and the length
of the beam. The beam length is increased from 6 m to 12 m
tf 2.0 in increments of 1.0 m. Thus, seven different lengths are used
-.: for each given W shape and the total number of training
i8 1.5 instances is 196.
For testing the network, 10 different W shapes from the
j 1.0 AISC ASD specifications (AISC 1989) are used (Table 4).
~ For each untrained instance, the beam length is changed from
E0.5 6.5 m to 11.5 m in increments of 1.0 m. Thus, the trained
Downloaded from ascelibrary.org by Technische Universitat Munchen on 07/04/15. Copyright ASCE. For personal use only; all rights reserved.

~ network was tested using 60 untrained instances. The desired


~ 0.0 '---:'-:--~-----''-----'----'---'---'-----'------'-,-----' outputs consist of two elastic critical lateral torsional moments
-1.0 -0.8 -0.6 -0.4 -0.2 0.0 0.2 0.4 0.6 0.8 1.0
End moments ratio (~)
corresponding to the simply supported beam (M erl ) and fixed-
end beam (M"'2) for the given length (L). They are computed
FIG. 8(a). Desired Moment-Gradient Coefficient Value by Energy from the closed-form solutions (Chen and Lui 1987)
Method

~
3.5 r - - - - - - - - - - - - - - - - - - - - - ,
(b)
2
+7f- - -
GlU
EC"'J (9)
o 3.0 ~ Eq. 12
~ -
.........
set-1
set-2
/'/'
/'
2
+ 47f EC",]
tf 2.5 ~.-:::..-::..-:. ,-_ _- . (10)
-- set-3 GlU
-....
·i 2.0
- -- set-4 ~~.~

.,",,'
.,/..,.-
.... ......-:

where E = modulus of elasticity of steel; I,. = moment of


~_c: 1.5 inertia about the minor axis; G = shear modulus of elasticity
of steel; 1 = torsional constant; and C", = warping constant.
~ For this example, the number of nodes in the input layer is
e 1.0
~ three, in the competition layer 196 (equal to the number of
~ 0.5 training instances), and in the interpolation layer two (Fig.
o
::E 5). This results in a topology with 1,176 links, a rather large

l:!
w
g 10
o '-----'----'---"--~--"---'-!---'-.....-
-1.0 -0.8 -0.6 -0.4 -0.2 0.0 0.2 0.4 0.6
......- ' " -
0.8
] 1.0
network. Learning results for the untrained instances are shown
in Fig. 6. The number of winning nodes in competition layer
was set to four and used for calculating the output of testing
instances. For both cases of simply supported and fixed-end
beams, the maximum error in critical moments for all 60
End moments ratio (J})
verification instances is found to be 2.66% and 2.46%, re-
FIG. 8(b). Prediction of Moment-Gradient Coefficient by CPN spectively. These are within the acceptable limits in design
computations.
cates clearly that the maximum percentage error of bending
moments in both directions for untrained instances is much Moment-Gradient Coefficient Cb for Doubly and Singly
lower for the counterpropagation network. Training the coun- Symmetric Steel Beams Subjected to End Moments
terpropagation network took 0.029 CPU seconds on Cray
YMP8/864 and approximately 3 sec on a 486-based micro- The maximum buckling moment, Men in singly symmetric
computer. According to Vanluchene and Sun (1990), training steel beam under nonuniform moments can be approximated
the back-propagation network required approximately 32 hr by using the moment-gradient coefficient, Ch , as follows (AISC
of CPU time on a personal computer (the type of the pro- 1989)
cessor was not specified). (11 )

APPLICATION TO LARGE NETWORKS where Mum = critical lateral torsional buckling moment of
simply supported beams under uniform moment. The coef-
The two examples presented in the previous sections used ficient Ch can be determined only approximately. AISC ASD
small networks. In this section, we apply the counterpropa- specifications (AISC 1989) recommends the following equa-
gation neural network to two structural engineering problems tion:
that require large networks.
C h = 1.75 + 1.0513 + 0.3(3" :s 2.3 (12)
Lateral Torsional Buckling of Steel Beams
where 13 = M] /M 2 = ratio of the two end moments, M 1 being
This example is the prediction of elastic critical lateral tor- the smaller of the two. This ratio is defined as positive when
sional buckling moments of wide-flange steel beams (W shapes) the end moments cause double-curvature deflection, and neg-
subjected to a uniform bending moment (Fig. 5). Two dif- ative otherwise. The popularity of this equation is due to its
ferent boundary conditions are considered: (1) Simple sup- simplicity only. This equation is a reasonable approximation
ports where beam ends are restrained against twisting and for doubly symmetric sections. However, it can be unsafe or
lateral translation but are free to rotate and warp; and (2) overly conservative for monosymmetric beams such as T-sec-
fixed supports where beam ends are restrained against lateral tions, especially under high moment gradient (Kitipornchai
displacement and warping but are free to rotate about the et al. 1986). Kitipornchai et al. (1986) proposed the following
horizontal axis. Twenty-eight different W-shapes from the complicated equation instead of (12):
1210 I JOURNAL OF STRUCTURAL ENGINEERING I AUGUST 1995

J. Struct. Eng. 1995.121:1205-1212.


. A pro~lem with the back-propagation algorithm is the ar-
C = Mum
p
M", = Y.5.'IT [J1 + KZ + ('ITS)Z + 'ITS] -[
2 2
(13) blt~ary t~la!-and-error selection of learning and momentum
r~tlos. SImIlarly, we have the arbitrary trial-and-error selec-

where 'Yc = nondimensional elastic critical buckling moment· tIon of th~ learning coefficient a in the competition layer and
K = Y'IT 2ECw IGJL z ; l) = (R>xIL)YElyIGJ; and R>x = mon~ the lear~Ing coefficient b in the interpolation layer for the
osymmetry parameter (R>x = 0 for doubly symmetric beams). conventIOnal counterpropagation algorithm. In this research,
The nondimensional elastic critical buckling moment 'Y is we circ.umven!ed t~is problem by using a simple formula as
~ompute~ by the Rayleigh-Ritz method. In this pap~r,c'we a functIOn of IteratIon number for both coefficients and ob-
tained excellent convergence.
mclude nme terms of the Fourier sine series as the buckled
shape function for both lateral deflection and rotation about The counterpropagation algorithm was compared with the
back-pro~agation. algorithm using two examples previously
the longitudinal axis in order to calculate the desired output.
reported In the hterature. We found superior convergence
Thus, the counterpropagation network (Fig. 7) is trained
to predict Cp for monosymmetric beams under nonuniform property for the counterpropagation network and a substan-
Downloaded from ascelibrary.org by Technische Universitat Munchen on 07/04/15. Copyright ASCE. For personal use only; all rights reserved.

bending. Each training instance for learning this example con- tial decrease in the CPU time. In one example for which the
sists of four nondimensional parameters: the ratio of the mo- exact data were available, we found the counterpropagation
ments of inertia of the compression flange and the whole cross algorithm to be 3,000 times faster. The reason for CPN's
sectio_n with respect to the minor axis (p), the beam param- better convergence property over BPN's is not entirely due
eter K = Y'IT zElvh zI4GJU, where h is the distance between our definition of learning coefficient, (4). The superior con-
flange centroids,' the ratio of the moments of inertia of the vergence of CPN is also due to the fact that it is less sensitive
section about the major and minor axes (I)Ix ), and the end to the learning coefficients. In BPN, for a given training in-
moment ratio (R». To cover most types of monosymmetric stance the weight-adjustment process occurs across all con-
beams, from inverted T-shape (p = 0) to T-shape (p = 1.0), nection weights for given learning coefficients. In CPN, how-
the following combinations of the four parameters are con- ever, the weight adjustment process for a given training instance
sidered as training instances: affects only the connection weights associated with the win-
ning node. In BPN, during a training process, a large number
k: {D. I, 0.5, 1.0, 1.5, 2.0} of iterations is required for the connection weights to stabi-
p: {O.O, 0.2, 004, 0.6, 0.8, 1.0} liz.e .. F~r.each training instance, a new set of connection weights
mInImIZIng the system error must be calculated. In CPN, on
/,//,: {0.005, 0.01} the other hand, for each training instance, only the specific
connection weights associated with the nodes are to be ad-
13: {-1.0, -0.8, -0.6, -004, -0.2,0.0,0.2, justed.
004, 0.6, 0.8, 1.0} In addition to the two small neural network examples re-
ported in recent structural engineering literature, we created
Thus, the number of training instances is equal to 528. De- two large examples, with 1,176 and 4,224 links and 196 and
sired output consists of the two moment-gradient coefficients 528 training instances, respectively. The second example is a
defined by (12) and (13). The number of nodes in the input complex stability analysis problem requiring extensive nu-
layer is four, in the competition layer 528, in the interpolation merical analysis and management of a large amount of data.
layer two, and the number of links in the network is 4,224. In each case, the counterpropagation network was trained in
The number of winning nodes in the competition layer is set less than 30 iterations in both competition and interpolation
to three for this problem. Forty-four untrained instances were lay~rs. The verification instances proved the counterpropa-
used to test the performance of the network. For all the gatIon neural network can produce results within a few per-
verification instances, the coefficient Cb defined by (12) is centage of the exact or desirable values and reasonable amount
predicted with an error of less than 1%. Thus, the results of of CPU time. Of course, the system error can always be
learning corresponding to (12) are not presented. further reduced by additional training.
. The results of learning C b corresponding to (13) are given In a back-propagation neural network, the solution is based
III Table 5. As the end moment ratio (R» increases, the max-
on the minimization of the system error. The stabilized weights
imum percentage error for the Cb corresponding to (13) gen-
do not guarantee a global minimum for the system error. In
e~ally mcre.a~es up to 7.1 %. Prediction of the moment-gra-
dient coeffICIent Cb for monosymmetric steel beams under other words, the back-propagation neural network is equiv-
n?nuniform bending by the counterpropagation is plotted in alent to an unconstrained optimization problem. In the coun-
FIg. 8. The 44 verification instances are divided into four sets terpropagation algorithm, on the other hand, the specific con-
depending on the value of the parameter K (0.62, 0.74, 1.34, nection weights associated with a winning node, instead of
and 1.87). It took 2.77 CPU seconds for training the network all connection weights, are adjusted to minimize the error for
using 528 training instances and testing 44 verification in- a given training instance. The learned output from a coun-
stances, where it required 8.09 CPU seconds to determine terpropagation network is simply the best result from the
572 nondimensional elastic critical buckling moments, 'Yn as encoded information based on the given training instances.
training and verification instances by conventional analytical As a side result of this research, we found that for mon-
methods on Cray YMP8/864. osymmetric thin-walled sections the simple equation in the
AISC ASD specification for the moment-gradient coefficient,
CONCLUSIONS C b , is inadequate for moment ratios greater than 0.2. Gen-
erally, this error increases with an increase in the moment
The back-propagation neural network algorithm has re- ratio. While the error is often on the conservative side for
ce~tly been applied to structural engineering examples. In many cases, the error is on the unconservative side for largely
thiS paper, counterpropagation neural networks with com- monosymmetric beams.
petition and interpolation layers was presented and applied
to several structural engineering examples. The algorithm was ACKNOWLEDGMENTS
developed in FORTRAN on a Cray YMP8/864 using a CF77
compiler. This paper is based upon work supported by the National Science
Foundation (Grant No. MSS-92221l4). the American Iron and Steel

JOURNAL OF STRUCTURAL ENGINEERING / AUGUST 1995/1211

J. Struct. Eng. 1995.121:1205-1212.


Institute, and the American Institute of Steel Construction. Computing neural network applications using backpropagation." Microcomputers
time was provided by the Ohio Supercomputing Center. in Civ. Engrg., 9(2), 145 - 159.
Hung, S. L., and Adeli, H. (1991a). "A model of perceptron learning
APPENDIX. REFERENCES with a hidden layer for engineering design." Neurocompu,ting. 3(1).
3-14.
Adeli, H. (1992). Parallel processing in computational mechanics. Marcel Hung, S. L., and Adeli, H. (1991b). "A hybrid learning algorithm for
Dekker, New York, N.Y. distributed memory multicomputers." Heuristics, 4(4), 58 - 68.
Adeli, H. (1992). Supercomputing in engineering analysis. Marcel-Dek- Hung, S. L., and Adeli, H. (1993). "Parallel backpropagation learning
ker, New York, N.Y. algorithm on Cray YMP8/864 supercomputer." Neurocomputing. Vol.
Adeli, H., and Hung, S. L. (1993a). "A concurrent adaptive conjugate 5, 287 - 302.
gradient learning algorithm on MIMD shared-memory machines." J. Hung, S. L.. and Adeli, H. (1994a). "Object-oriented backpropagation
Supercomputer Appl., 7(2), 155 - 166. and its applications to structural design." Neurocomputing, 6(1). 45-
Adeli, H., and Hung, S. L. (1993b). "A fuzzy neural network learning 55.
model for image recognition." Integrated Computer-Aided Engrg., 1(1), Hung, S. L., and Adeli, H. (1994b). "A parallel genetic, neural network
43- 55. learning algorithm for MIMD shared memory machines." IEEE Trans.
Adeli, H., and Hung, S. L. (1994). "An adaptive conjugate gradient
Downloaded from ascelibrary.org by Technische Universitat Munchen on 07/04/15. Copyright ASCE. For personal use only; all rights reserved.

on Neural Networks.
learning algorithm for effective training of multilayer neural net- Hurson, A. R., Pakzad, S., and Jin, B. (1994). "Automated knowledge
works." Appl. Mathematics and Computation, 62(1), 81-102. acquisition in a neural network-based decision support system for in-
Adeli, H., and Yeh, C. (1989). "Perceptron learning in engineering complete database systems." Microcomputers in Civ. Engrg., 9(2),
design." Microcomputers in Civ. Engrg., 4(4), 247-256. 129 - 143.
Adeli, H., and Yu, G. (1993). "An object-oriented data management Kang, H.-T., and Yoon, C. J. (1994). "Neural network approaches to
model for numerical analysis in engineering." Microcomputers in Civ. aid simple truss design problems." Microcomputers in Civ. Engrg..
Engrg., 8(3),199-209. 9(3).211- 218.
AdelL H., and Zhang, J. (1993). "An improved perceptron learning Kitipornchai, S., Wang, C. H .. and Trahair. N. S. (1986). "Buckling of
algorithm." Neural, parallel, and scientific computations, 1(2),141- monosymmetric I-beams under moment gradient." J. Struct. Engrg..
152. ASCE, 112(4), 781-799.
AISC. (1989). Manual of steel construction, allowable stress design. Chi- Kohonen, T. (1988). Self-organization and associative memory. Springer-
cago, Ill. Verlag, New York, N.Y.
Chen, W. F., and Lui, E. M. (1987). Structural stability. Elsevier, New Masri, S. F., Chassiakos, A. G., and Caughey, T. K. (1993). "Identi-
York, N.Y. fication of nonlinear dynamics system using neural networks." J. Appl.
Elkordy, M. F., Chang, K. c., and Lee, G. C. (1994). "A structural
Mech., 60(1),123-133.
damage neural network monitoring system." Microcomputers in Civ.
Messner, J. I., Sanvido, V. E .. and Kumara, S. R. T. (1994). "StruetNet:
Engrg., 9(2). 83 - 96.
a neural network for structural system selection." Microcomputers in
Ghaboussi, J., Garrett, J. H., and Wu, X. (1991). "Knowledge-based
Civ. Engrg., 9(2), 109 - 119.
modeling of material behavior and neural networks." 1. Engrg. Mech.,
117(1),132-153. Rumelhart, D., Hinton. G., and Williams. R. (191\6). "Learning rep-
Grossberg. S. (1982). Studies of mind and brain. Reidel Press, Boston, resentations by back-propagation errors." Parallel distributed pro-
Mass. cessing; Vol. 1, D. Rumelhart et al.. eds., MIT Press. Cambridge,
Gunaratnam, D. J., and Gero, J. S. (1994). "Effect of representation Mass., 318-362.
on the performance of neural networks in structural engineering ap- Saleh, A., and Adcli, H. (1994). "Microtasking, macrotasking, and au-
plications." Microcomputers in Civ. Engrg., 9(2), 97 - 108. totasking for structural optimization." 1. Aerosp. Engrg.. ASCE, 7(2).
Hajcla, P.. and Berke, L. (1991). "Neurobiological computational modes 156-174.
in structural analysis and design." Computers and Struct., 41(4), 657- Stephens, J. E .. and Vanluehene, R. D. (1994). "Integrated assessment
667. of seismic damage in structures." Microcomputers in Civ. Engrg.. 9(2).
Hecht-Nielsen, R. (1987a). "Counterpropagation networks." Proc., IEEE 119-128.
1st Int. Conf. on Neural Networks; Vol. 11, IEEE Press, New York, Vanluchene, D., and Sun, R. (1990). "Neural networks in structural
N.Y., 19-32. engineering." Microcomputers in Civ. Engrg., 5(3). 207 - 215.
Hecht-Neilsen, R. (1987b). "Counter propagation networks." Appl. Op- Woods, D. (1988). "Back and counter propagation abberations." Proc.,
tics, 26(23). 4979 - 4985. IEEE Int. ConI Oil Neural Networks, Vol. I. IEEE New York, N.Y ..
Hecht-Neilsen, R. (191\1\). "Applications of counterpropagation net- 473-479.
works." Neural Networks, 1(2), 131- 139. Yu, G., and Adeli. H. (1993). "Object-oriented finite element analysis
Hegazy, T .. Fazio. P., and Mosclhi, O. (1994). "Developing practical using EER model." J. Struct. Engrg., ASCE, 119(9),2763-271\1.

1212/ JOURNAL OF STRUCTURAL ENGINEERING / AUGUST 1995

J. Struct. Eng. 1995.121:1205-1212.

You might also like