Professional Documents
Culture Documents
951810
416
Downloaded from SAE International by University of Edinburgh, Friday, August 10, 2018
data included in the database can be presented as a series seven-dimensional space. The relationships between the
of points (Fig. 3 - 5). By connecting those points, the dimensions and stress concentration factors are
stress concentration factor can be predicted for a shape interrelated and therefore cannot be handled easily.
not included in the database. Here, there are six For example, the relationship between the web
dimensions for the horizontal axis, so making an thickness and the stress concentration factor varies
approximation with the series of graphs is equivalent to widely according to the rotating radius of the crankshaft
making a function which shows curved surface in the (Fig. 3). Fig. 4 shows the relationship between the radius
of a crankpin fillet and the stress concentration factor,
where the stress concentration factor diverges as the
dimension approaches zero. On the other hand, in Fig. 5,
the stress concentration factor approaches a constant
value as the journal diameter increases.
From the above discussion, two conditions are found
to be necessary for the functional approximation:
1.It must be able to perform approximation by
multivariable functions.
2.It must be able to generate nonlinear functions
which can correctly express trends such as
divergence and saturation.
FUNCTIONAL APPROXIMATION WITH THE
ARTIFICIAL NEURAL NETWORK - Fig. 6 shows a
conceptual drawing of a general hierarchical artificial
neural network. The solid circles in the figure represent
neurons, and these are interconnected to make up a
network. Such a neural network would behave as a
multivariable function. That is, if a set of numerical
values is input into each neuron of the first layer, each
neuron calculates its input value to obtain the output
value in accordance with a determined rule, and one
numeric value is output from the right end. In this paper,
the dimensions of the crankshaft are given as the inputs,
and the stress concentration factor is obtained as the
output for use in the prediction of the stress in the
crankshaft.
The characteristics of the multivariable function
obtained from the network are dependent on the transfer
function and the numeric values (connection weights and
threshold values) defined for each neuron. The neural
network automatically sets those numeric values based on
input/output data, and performs calculations which
construct the relationship between the inputs and outputs
shown by the data on the network. These calculations
417
Downloaded from SAE International by University of Edinburgh, Friday, August 10, 2018
are called “learning” of an artificial neural network, and network. In this case, however, the conventional
correspond to approximations of data with multivariable technique cannot be used, since some diverging curves
functions. are also involved.
Various algorithms have been proposed for this To satisfy the conditions previously mentioned about
learning calculation, and the error back propagation is approximation, the adaptive transfer function algorithm,
one of the most common techniques[6]. In many cases, or adaptive neuron algorithm (APPENDIX (1)) has been
the saturation functions called sigmoid functions (Fig. 6) developed. This technique has two advantages:
are used as the transfer functions to correct the values 1.It can use composite transfer functions which
(connection weights and threshold values) in the combine saturating and/or diverging functions.
2.The characteristics of the transfer functions
change and fit various conditions by learning.
This technique allows the use of arbitrary transfer
functions based on objective analogy. As a result,
approximations with more various curves or curved
surfaces become possible than ever.
OFTHEFUNCTIONAL RESULTS
approximations is shown in
A AP ROXIMATIONS - One
7.
Fig. functional
the
of
result
close to the given data point. smooth curve passes
The structure of the network applied to
this problem is shown in Fig. 8. Composite type
transfer functions, which can be used for both diverging
and saturating cunes, were applied to the neurons in
the intermediate layer of the hierarchy. The types of
those functions and the structure of the network such as
number of neurons located in the intermediate layer were determined
in consideration of the precision of the ap roximations
and the shape of the curved surfaces. The
process of learning performed by the
artificial neural network is shown in Fig. 9.
The horizontal axis represents the number of repeated corrective calculations
(number of learning operations), while the vertical
axis represents the mean square error from each
data created after learning. The results of a total
of 7,400 repetitive calculations (learning operations) showed
a mean square error for each data of
5.9%. As already described, about 130 sets of data were
prepared, but it might seem that these were not enough to
create the approximate functions through interpolation into all
the combinations of input values. This will
be
considered separately (APPENDIX 2). 2). 2). 2). separateh'
(APPENDIX 2).
in the verification are not included in the database used calculated using the finite element method. Furthermore,
for the learning of the neural network. As shown in Fig. an adaptive transfer function algorithm was introduced
10, the stress concentration factors can be roughly as a learning method for the neural network. As a result,
divided into two groups of about 3.0 and 5.5. The former the creation of approximate functions suitable for the
represents the crankshafts of single-cylinder small utility relevant purpose from a limited number of data became
engines, and the latter represents the crankshafts of possible. By using this system, the stress concentration
3-cylinder outboard motors. factors in crankpin fillets can be predicted precisely in a
short time. The predicted values compared to the
CONCLUSION measured values fell in the range of -6.9 to +6.3%.
REFERENCES
419
Downloaded from SAE International by University of Edinburgh, Friday, August 10, 2018
Where � 3B1;, � 3B2; and � 3B3;m are small positive constants, called
“learning parameters.”
As an example, the transfer function used in the
neural network for stress prediction is shown as follows:
420
Downloaded from SAE International by University of Edinburgh, Friday, August 10, 2018
function, km m � 3D; 1 to 4) represents the characteristic and prediction is lowered in conventional artificial neural
value of the function. That is, k1 and k2 are the gradients networks.
of the linear and parabolic functions, and k4 is the One method to improve precision is to calculate the
gradient of the sigmoid function. The magnification stress concentration factors with all variables
range of the sigmoid function is represented by k3. Fig. interpolated. If there are six input variables, each of
14 shows the behavior of the function according to which requires four data to indicate its overall trend, the
variation of these characteristic values. number of required data to be interpolated is 46 � 3D; 4,096.
With the adaptive transfer function algorithm, any The precision of both approximation and prediction can
differentiable function such as that in the EQ(5) can be be satisfied by increasing the number of interpolated data,
used as the transfer function, and any parameter which but such an increase is not practical for the amount of
determines the characteristics of the transfer function can calculations required to create data.
be learned. Furthermore, the ability of this artificial The other way is to use the composite transfer
neural network to optimize the transfer function functions in the artificial neural network and to learn the
characteristics enables learning with fewer trials and less database maintaining the trend of the stress
error. concentration factor for dimensional values even in areas
not interpolated. A graphical monitoring system was
(2)INSIGHT INTO LEARNING BY AN ARTIFICIAL used to observe the curve shape of the stress
NEURAL NETWORK concentration factor and this made the learning process
The approximate functions created by an artificial more effective. Furthermore, the network comprises the
neural network have some errors compared to the learned least number of required neurons to prevent over-
data, and are not always located on the series of original learning.
data points. Therefore, the precision of functional
approximation by an artificial neural network and the
precision of the prediction with the approximate function
need to be considered separately. The former is based on
the error occurring in the approximation according to the
values in the database, and the latter is based on the error
in predicting the values to be obtained.
For example, if a small number of learning data are
learned by a large network with many neurons, the
approximation error can be zero. However its
approximated cune often twists between the points and
has meaningless points of inflection. This is called
“over-learning” of a neural network (Fig. 15). The same
behavior is seen when data are interpolated with a
high-degree polynomial. The over-learned function has
good approximate precision, but cannot reflect the
overall trend of the data series, resulting in low
prediction precision both in interpolated and extrapolated
areas. If the number of neurons are decreased to prevent
over-learning, the precision of both the approximation
421