You are on page 1of 8

The use of Kernel PCA in Evolutionary

Optimization for Computationally Demanding


Engineering Applications
Dimitrios Kapsoulis, Konstantinos Tsiakas, Varvara Asouti, Kyriakos Giannakoglou
Parallel CFD & Optimization Unit
School of Mechanical Engineering
National Technical University of Athens
Athens, Greece
Email: {jim.kapsoulis, tsiakost}@gmail.com, vasouti@mail.ntua.gr, kgianna@central.ntua.gr

AbstractTwo techniques to further enhance the efficiency of [4] and Gaussian processes [3] are among the most widely
Evolutionary Algorithms (EAs), even those which have already used metamodels. In this work, without loss in generality,
been accelerated by implementing surrogate evaluation models Radial Basis Function (RBF, [4]) networks, which are three
or metamodels to overcome a great amount of costly evaluations,
are presented. Both rely upon the use of a Kernel Principal layer ANNs, are used as metamodels. With the exception of
Component Analysis (Kernel PCA or KPCA) of the design the few starting generations of the MAEA, all population
space, as this reflects upon the offspring population in each members are approximately evaluated on local metamodels
generation. The PCA determines a feature space where the trained on the fly (online metamodels; to be explained in
evolution operators should preferably be applied. In addition, Section II) and only the most promising among them undergo
in MetamodelAssisted EA (MAEAs), the PCA can reduce the
number of sensory units of metamodels. Due to the latter, the reevaluation on the PSM. The inexact preevaluation (IPE)
metamodels are trained at much lower computational cost while of the population members in each generation is a key feature
approximating better the objective function value. This paper of our method, see [1], [3], [5]. This MAEA differs noticeably
extends previous work by the authors which was based on from the majority of other existing MAEAs which rely upon
Linear PCA, used for the same purposes. In the present paper, a single offline metamodel trained during an expensive pre
the superiority of using the Kernel (rather than the Linear)
PCA, especially in realworld applications, is demonstrated. The processing phase based on Design of Experiment strategies
proposed methods are assessed in single and twoobjective [6].
mathematical optimization problems and, finally, showcased in Realworld optimization problems may have a great number
aerodynamic shape optimization problems with computationally of design variables. The socalled curse of dimensionality
expensive evaluation software. refers to both the performance degradation of EAs or MAEAs
in the presence of a great number of design variables and
I. I NTRODUCTION
the fact that the prediction ability of metamodels becomes
During the last decades, EAs are widely used to solve questionable with many input units, for a reasonably low
Single and Multiobjective optimization problems (SOO or number of training patterns.
MOO) with or without constraints, by accommodating any Even without the use of metamodels (i.e. in standard EAs),
readytouse evaluation tool. Though EAs are able to find the search process becomes quite faster, if the objective
global optima, they require a great number of calls to the function f is separable. By definition, f is separable if there
problemspecific model (PSM). If the PSM is a computa- exist functions f1 , f2 , . . . , fN , such that: f (x1 , x2 , . . . , xN ) =
tionally demanding software, then the optimization turnaround f1 (x1 )f2 (x2 ) . . . fN (xN ). In such a case, it suffices to min-
time may become prohibitively large. Software pertinent to imize f1 (x1 ) in terms of x1 , f2 (x2 ) in terms of x2 and so
Computational Fluid Dynamics (CFD), Computational Struc- on. Since EAs perform better in separable problems, it would
tural Analysis, Computational Electromagnetics etc. are typi- be beneficial to transform/map the design space to a new one,
cal examples of PSMs used in realworld applications. usually referred to as the feature space, where the problem
Various methods capable of reducing the computational cost becomes as separable as possible. This mapping can be
of an EAbased optimization have been devised. In MAEAs carried out through the PCA of an appropriate data set.
[1], [2], [3], surrogate evaluation models (often referred to Moreover, artificially reducing the number of design vari-
as metamodels) can be used to replicate the outcome of ables the metamodel sees from N to N NDR , using a di-
the PSM at negligible CPU cost, after being trained on a mensionality reduction technique, may improve its prediction
set of already evaluated individuals. Nowadays, MAEAs have ability and reduce the corresponding training cost. This will
become a standard tool in evolutionary optimization. Response be the case, if the N NDR design variables the metamodel
surface models (RSM), artificial neural networks (ANNs) is trained on are the ones that affect mostly the objective
IP E
function. So, in each generation of a MAEA, during the IPE starts after a predefined minimum number (TDB ) of already
phase, the PCA reduces the number of features the ANN evaluated individuals is archived in the DB; up to this point,
sensory units see. In other words, the metamodels are trained a standard EA is used. This MAEA relies on online trained
on patterns having less components than the design variables metamodels; these are trained separately for each new popu-
array. Keeping only components having large variances, the lation member on its closest (in terms of Euclidean distances
IPE phase of the MAEA becomes much more effective. in the normalized design space) already evaluated individuals
The use of PCA for improving the performance of EAs found in the DB, [5]. The selection of RBF centers and training
and/or MAEAs, in large scale problems with a great number patterns is important since it affects the prediction ability of the
of design variables, has been proposed in [7]. The PCA assists metamodel. At the end of the IPE phase, just a few (e )
the application of evolution operators and the training of meta- top population members, i.e. the most promising ones based
models. In the present work, the Kernel PCA is implemented on the metamodel, undergo reevaluation on the PSM.
instead of the Linear PCA used in [7]. It will convincingly be
demonstrated that this boosts further the convergence of the III. PCA- DRIVEN EA S AND MAEA S
EAs or MAEAs. Principal Component Analysis (PCA) uses an orthogonal
The first part of the results section is dealing with some transformation to convert a set (X) of observations x RN
widely used mathematical optimization problems. Due to their of possibly correlated variables into a set of uncorrelated
low CPU cost, these can be solved several times, for different variables called principal components. In our case, the data
random number generator (RNG) seeds and safe comparisons set X must be formed by collecting M possible solutions
can be reached. In the second part, the PCA variants of EA to the problem with some common characteristics. The PCA
and MAEA are used to solve aerodynamic shape optimization does not need responses to be known (unsupervised learning).
problems. In the (, ) EA or MAEA we are dealing with, X can be
formed by either the offspring of the last generation (their
II. T HE (, ) EA & MAEA common characteristic being that they gradually tend towards
Assume that EAs are used to solve the following optimiza- the optimal solution(s), at least after the first explorative gen-
tion problem with Mo objective functions and Mc constraints erations) or the current elite set. The transformation based on
the outcome of the PCA correlates each principal component
minF (x) = min{f1 (x), ..., fMo (x)} with a variance of the observations; the first component is
subject to cj (x) 0, j = 1, . . . , Mc (1) in the direction of the largest variance, the second is in the
direction of the second largest variance and so forth.
where x RN is the optimization (or design) variables array. Assume that the simplest possible variant of PCA (Linear
For all design variables, their lower and upper bounds are PCA) is implemented on a set of observations. It is assumed
known. For Mo>1, eq. 1 represents a MOO problem; in such that the data set X has been postprocessed so as to have zero
a case, the EA computes the front of nondominated solutions mean (E[X] = 0) and unit standard deviation (E[X2 ] = 1) along
(Pareto front). all directions. Its covariance matrix is PN N = M 1
XXT . This
A brief description of the background (, ) EA used in this matrix contains the observations correlation and is decom-
paper follows. It handles three populations, namely the parent posed as PN N = U U T where is a diagonal matrix with
(Sg , with members), the offspring (Sg , with members) the eigenvalues of P and U is the matrix with the eigenvectors
and the elite one (Sg , with members at most), where g is of P as rows. These eigenvectors are the principal components
the generation counter. Sg results from the application of the and the corresponding eigenvalues are their variances. Thus,
parent selection operator to (Sg1 Sg1 ). Sg results from a feature space based on the principal components becomes
Sg , via crossover and mutation, including elitism. In MOO available.
problems, the Pareto front is approximated by the S of the last Linear PCA performs well in problems with linearly cor-
generation. Each offspring is given a scalar cost by processing related variables [7], [11]. However, in complex nonlinear
the fi values of the current population based on dominance problems, Linear PCA may not perform well and Kernel PCA
and sharing criteria such as those involved in the widely (KPCA) [12] is, indeed, a viable alternative.
used SPEA and NSGA techniques, [8], [9]. In constrained KPCA transforms the design space to the feature space
problems, individuals violating one or more constraints are pe- through a mapping function, : RN RL . L can be arbitrarily
nalized usually with an exponential penalty function. In SOO, large. Also, as it will be shown below, there is no need to
the scalar cost is nothing more than the objective functions explicitly define the mapping function . Given a data set X,
value. All the already evaluated individuals, paired with the the covariance matrix PLL (or just P hereafter) elements are
corresponding objective function and constraint values, enter
1
M
a database (DB) which is dynamically updated during the
evolution; duplicate entries are not allowed. P = (xi ) (xi ), , = 1, . . . , L (2)
M i=1
In the MAEA, ANNs or any regressor may serve as meta-
models, undertaking the IPE of the newly formed offspring The corresponding eigenproblem is expressed by the fol-
in each generation [1], [10]. The IPE of the Sg members lowing system of L equations
2) Compute the eigenvalues () and eigenvectors (a) via
Pv r = r v r , r = 1, . . . , L (3) the solution of eqs. 10.
3) Each new individual x is projected, into the feature space
where is the (LL) diagonal matrix with the corresponding as follows
eigenvalues. Depending on the value of L, the solution of eqs.
3 may become prohibitively expensive. The socalled kernel
M
x) =
Cr (x) = v r ( ari k(xi , x), r = 1, . . . , M
trick can help overcome this problem. Each eigenvector can
i=1
be written in the form (11)

M In the present method, transforming the individuals back to
v r = x ), r = 1, . . . , L
ar ( (4) the design space is absolutely necessary. In Linear PCA this
=1 is straightforward, which is not the case for the Kernel PCA
x ) RL might not have
though. The reason is that a vector (
or, after plugging eqs. 4 into 3,
a unique representation in R . Nevertheless, a vector z RN
N

which approximately maps to ( x ) can be found. The design

L
M
M

vector x R maps to C(
N
x ) RM as follows
Pij ar j (x ) = r ar i (x ), i = 1, . . . , L (5)
j=1 =1 =1

M
xq ), q = 1, . . . , M , yields
The inner product of eqs. 5 and ( Cr (x ) = ari k(xi , x ), r = 1, . . . , M (12)
i=1
a system of M equations,
This representation compresses the information of the fea-

L
L
M ture space into M dimensions, while avoiding to compute the
ar Pij i (xq )j (x ) = mapping function . A projection of C( x ) to the RL space
i=1 j=1 =1 is defined as

M

L
M
p(x ) = Cr (x )v r (13)
r ar i (x )i (xq ), q = 1, . . . , M (6)
r=1
i=1 =1

Using eq. 2, eqs. 6 can be written as The minimization of the Euclidean distances between the
projection p(x ) and (
z ) leads to the computation of


L
L
M
M p(x ) (
z = argmin|| z ))|| (14)
ar j (xz )i (xz )i (xq )j (x ) =
i=1 j=1 =1 z=1 If the RBF kernel, eq. 9, is used, problem (14) can be solved

L
M via the following fixed point iterative algorithm, [13],
M r ar i (x )i (xq ), q = 1, . . . , M (7)
i=1 =1 M M ||
xj z old ||22
new i=1 x )aij exp(
j=1 Ci ( 2 2 )xj
The (M M ) kernel matrix K is defined as z = M M ||
xj z old ||22
(15)
i=1 j=1 Ci (x )aij exp( 2 2 )

L Having presented the PCA method, its integration in EA
xi )
Kij = k(xi , xj ) = ( T (xj ) = p (xi )p (xj ) (8) based optimization follows. The PCA can be used during the
p=1 evolution, to reduce the turnaround time of the optimization,
Among the most widely used kernels are the polynomial, by transforming the nonseparable problem to an almost
the radial basis function (RBF) and the sigmoid ones [12], separable one. This transformation can also be used during
[13]. In this paper, the RBF kernel, the IPE phase, by training metamodels with less input units.
( )
||xi xj ||22 A. EA with PCAdriven evolution operators
k(x , x ) = exp
i j
(9)
2 2 In an EA assisted by the PCA, the role of the latter is
where is the width constant, is exclusively used. to identify the principal component directions and map the
Combining eqs. 7 and 8, the final system of equations to be design space to a new feature space, in which the problem
solved (instead of eq. 3) for each eigenvalue is becomes more separable. For the first gP CA generations,
the PCA is not applied. In all subsequent generations, the
Kaq = M qaq , q = 1, . . . , M (10) current offspring population is considered to be the set of M
training patterns. The offspring population is used due to its
where aq is the corresponding eigenvector. The use of Kernel diversity (in the course of generations, it becomes populated
PCA within an EA or MAEA includes the following steps: by members which, or the majority of which at least, are
1) Compute the kernel matrix (eq. 8) using the current close to the optimal solutions) and because it includes vital
offspring population. information for the sought optimal solutions. Once the design
space is mapped to the feature space via eq. 12, the evolution depending on whether the Linear or Kernel PCA is used. It is
operators are applied. evident that M(K)AEA(K) denotes a MAEA with a dual use
Regarding mutation, an increased mutation probability of Kernel PCA.
along the directions with small variances should preferably
IV. A PPLICATIONS T O M ATHEMATICAL C ASES
be used in order to reenforce the exploration in this direc-
tion. The mutation probability for each principal component A first demonstration of the proposed methods in four math-
(superscript i = 1, . . . , M ) is given by, [11], ematical problems, each of which with different characteristics
yi follows. The optimization platform EASY [14], developed by
pimut = pmut + (1 )M pmut M (16) the group of authors, was used to implement the PCA-based
i=1 yi variants of the (, ) EA and MAEA.
where pmut is a constant, userdefined, mutation probability, Three of them are SOO problems while the last one has two
[0, 1] (preferably kept below 0.5 to lay emphasis on the objectives. In all cases, the number of optimization variables
second term in eq. 16) and yi is computed as x = (x1 , . . . , xN ) is N = 50 and o RN is the optimal solution
Vmax Vi (known beforehand, in all of them). Let z =x o.
yi = (17)
Vmax Vmin 1) Ellipsoid Problem :
where Vi are the variances of the current offspring popula-
tion, after being mapped to the feature space and Vmax =
N

max{V1 , ..., VM }, Vmin = min{V1 , ..., VM }. min f (x) = zi2 (18)


After the application of the crossover and mutation op- i=1

erators, the population is transformed back to the design where 100 6 xi 6 100 and oi = i/N 2 . This function is
space using eq. 15. An EA or MAEA in which the evolution unimodal and separable.
operators are assisted by the Linear PCA will be referred to 2) Weierstrass Problem :
as an EA(L) or MAEA(L). In contrast, EA(K) or MAEA(K)
denote the use of Kernel PCA, for the same purpose.
N 20

B. EAs with PCATruncated Metamodels min f (x) = ( [ak cos(2bk (zi + 0.5))])
i=1 k=0
The computational cost of the metamodel training and its

20
prediction accuracy are affected by the problem size. Since N [ak cos(bk )] (19)
metamodels should be trained per generation, the training k=0
cost should be negligible compared to the cost of a single
where a = 0.5, b = 3, 0.5 6 xi 6 0.5 and oi = i/N .
evaluation on the PSM. Here, dimensionality reduction means
This function has several local optima with a single
to train the metamodels with patterns with less components
global optimum though. It is continuous over the whole
than the number of design variables. This noticeably improves
search space but differentiable only over a subset of it.
the prediction accuracy of metamodels.
Moreover, it is nonseparable, which makes it a nice
As mentioned before, the eigenvectors computed by the
testbed for assessing the PCA variants of EA.
PCA are associated with the corresponding eigenvalues, which
3) Rosenbrock Problem :
represent the variances. The variances indicate the directions
of the feature space along which the members of the cur-

N 1
rent generation are less or more scattered (a high variance
indicates that the available data are adequately scattered in min f (x) = (100(zi2 zi+1 )2 + (zi 1)2 ) (20)
i=1
this direction). The directions of the feature space with low
variances are ignored, since all members are so clustered where 0.5 6 xi 6 0.5 and oi = i/N . This function
around the same value in this direction, thus the corresponding is nonseparable, having a very narrow valley (in the
components of the training patterns are not expected to affect search space) separating local and global optima.
the prediction ability of the metamodel. The number (NDR )
of design variables to be cutoff is user-defined. Thus, the
metamodel is trained on data of lower dimension.
This truncation applies only during the IPE phase of
P CA
MAEAs and after TDB individuals have been archived in
the DB. In particular, for each population member, a person-
alized metamodel is built by selecting the necessary training
patterns as described in [5]. The training patterns are mapped
to the feature space and their principal components associated
with the lower eigenvalues (smaller variances) are truncated. A
MAEA in which the PCA is employed during the training of
the metamodels will be referred to as M(L)AEA or M(K)AEA
4) Two Objective Problem :
-110
EA
-120 EA(K)

N 1 -130 EA(L)

Objective Function
min f1 (x) = (100(zi2 zi+1 )2 + (zi 1)2 ) (21) -140
MAEA
MAEA(K)
i=1 M(K)AEA(K)
-150

N
20
-160
k k
min f2 (x) = ( [a cos(2b (wi + 0.5))])
i=1 k=0
-170

20 -180
N [ak cos(2bk 0.5)] (22) -190
k=0 -200
0 10 20 30 40 50
where a = 0.5, b = 3, 0.5 6 xi 6 0.5, z = x o1 , Evaluations/1000
=x o2 and o1i = i/N, o2i = i/N 2 .
w
Fig. 2. Weierstrass Problem: Average (of 20 runs) performance of the EA
In general, each mathematical case was optimized with and MAEA variants.
all variants of EA, namely EA, EA(L), EA(K), MAEA,
MAEA(K), M(K)AEA(K). Each optimization was repeated
20 times with different RNG seeds, so as to reduce their 14
EA
effect upon the optimization procedure. The convergence plots 12 EA(K)

Objective Function/106
present the averaged performance of the evolution in the EA(L)
10 MAEA
course of evaluations on the PSM of the 20 runs. For each MAEA(K)
8
run, a stopping criterion of 50000 evaluations was imposed. M(K)AEA(K)
A (10, 20) EA or MAEA was used in all cases. In the MAEA 6
IP E
variants, TDB = 100 and e = 3. In the runs assisted by the 4
P CA
PCA, gP CA = 2 and, when metamodels are used, TDB = 100,
2
NDR = 40.
Comparison of the convergence histories of all variants are 0
presented in figs. 1 to 4. For the twoobjective problem, the 0 10 20 30 40 50
hypervolume indicator measures the quality of the fronts of Evaluations/1000
nondominated solutions. This practically quantifies the area
dominated by the front of nondominated individuals up to Fig. 3. Rosenbrock Problem: Average (of 20 runs) performance of the EA
a userdefined nadir point; higher values correspond to better and MAEA variants.
fronts, i.e. those being closer to the Pareto front.
All PCA variants lead to faster convergence compared to 0.35
the corresponding EA and MAEA (the EA is compared to
Hypervolume Indicator

EA(L), (K) and the MAEA to MAEA(K) and M(K)AEA(K)). 0.3


The Kernel PCA constantly outperforms the Linear one.
0.25

14 EA
EA 0.2 EA(K)
Objective Function/10000

12 EA(K) EA(L)
EA(L) 0.15 MAEA
10 MAEA MAEA(K)
8 MAEA(K) M(K)AEA(K)
M(K)AEA(K) 0.1
6 0 10 20 30 40 50
4 Evaluations/1000

2 Fig. 4. Twoobjective Problem: Average (of 20 runs) performance of the EA


and MAEA variants.
0
0 10 20 30 40 50
Evaluations/1000 V. A ERODYNAMIC S HAPE O PTIMIZATION P ROBLEMS

Fig. 1. Ellipsoid Problem: Average (of 20 runs) performance of the EA and In this section, the aforementioned variants of EA and
MAEA variants. MAEA are used to optimize the shapes of an isolated airfoil
and a car, based on aerodynamic performance criteria.
A. Isolated Airfoil Shape Optimization M(K)AEA(K)) is presented in fig. 7 along with three indicative
airfoil shapes.
The first realworld optimization case deals with a two
objective problem in which the shape of an airfoil is optimized
for minimum drag (CD ) and maximum lift (CL ) coefficients.
The flow conditions are: freestream Mach number M = 0.19,
flow angle a = 5o and Reynolds number based on the chord
Rec = 1.3 107 .
The airfoil shape and the mesh deformations are controlled
by a morphing technique based on harmonic coordinates. This
is a generalization of barycentric coordinates and was initially
proposed in [15] for character articulation. A topologically
flexible structure called cage is used to control deformations.
Herein, this concept is extended to CFD applications where,
apart from performing the shape morphing, it is important to
simultaneously deform/displace the CFD mesh accordingly. To
avoid mesh quality degradation due to huge distortions at the
boundaries of the cage, a twocage structure is adopted instead
(fig. 5). Each cage is filled with a coarse unstructured mesh,
which is of course much coarser than the CFD mesh. The
harmonic coordinates of the vertices of the inner cage are Fig. 5. Shape Optimization of an Isolated Airfoil: The twocage structure
computed on the coarse mesh by solving as many Laplace used for shape parameterization and mesh deformation. The coordinates of
the vectices of the inner cage (red), excluding those of the leading and trailing
equations as the number of the inner cage boundary vertices, edge, stand for the design variables of the problem. The outer cage (blue) is
with appropriate boundary conditions. The socomputed har- not allowed to change. The (much finer) CFD mesh is not shown.
monic coordinates are, then, interpolated from the cage coarse
mesh to the CFD mesh nodes. A harmonic deformation of
the CFD mesh can, then, be explicitly defined by the cage
0.72
boundary vertices displacements. The inner cage controls the EA
0.7 EA(K)
shape deformation and mesh morphing while the outer cage
Hypervolume Indicator

limits the effect of morphing; in fact, the outer cage vertices 0.68 MAEA
MAEA(K)
remain still and the Laplace equations dont need to be solved 0.66 M(K)AEA(K)
for them. 0.64
In this case, 16 vertices were used to define the inner cage 0.62
(fig. 5). Two of them coincide with the leading and trailing 0.6
edge of the airfoil and remain still so as to keep the airfoil 0.58
chord constant. The remaining vertices were allowed to vary 0.56
along both directions, leading to 28 design variables in total. 0.54
The coarse cage mesh is formed by triangles generated by the 0 100 200 300 400 500
advancing front method. CFD Evaluations
The evaluation of each candidate solution is carried out on
the GPUenabled inhouse NavierStokes equations solver for Fig. 6. Shape Optimization of an Isolated Airfoil: Comparison of the
convergence history of the optimizations performed in terms of the number
compressible fluid flows [16], [17]. The CFD mesh is fully of CFD evaluations which is proportional to the CPU cost.
unstructured with 64 103 nodes. An averaged cost per
evaluation of a single individual on a single NVIDIA K20
GPU is 3min. This wall clock time includes the cost of B. Shape Optimization of the DrivAer Car
mesh deformation. The optimization was carried out using a The last case is concerned with the optimization of the shape
(20, 40) EA, EA(K), MAEA, MAEA(K) and M(K)AEA(K). In of the fastback configuration of the DrivAer car model. Air
IP E
the MAEA variants, TDB = 80 and e = 2. In the runs using flows in the axial direction with 11 m/s, while the road and the
P CA
the PCA, gP CA = 2 and, if metamodels are used, TDB = 80, cars wheels remain static. The aim is to redesign the car for
NDR = 18. A stopping criterion of 500 evaluations on the minimum drag coefficientz (CD ). The car shape parameteriza-
CFD model was imposed. tion was based on the volumetric NURBS method, [17], which
A comparison of the convergence histories of the EAs additionally undertakes the mesh deformation according to the
variants is presented in fig. 6. The integration of KPCA in EAs changing geometry of the rear part of the car. In particular, a
can compute a better front of nondominated solutions with 7 7 7 NURBS control box was used to parameterize the
the same computational budget. The optimal front of non volume around the cars boat tail and rear underbody, fig. 9.
dominated solutions from all the optimizations (computed by The three internal rows of control points in each direction
Fig. 9. Shape Optimization of the fastback DrivAer Car: NURBS morphing
box. Only the rear part of the car which is encapsulated into the morphing
box is allowed to change during the optimization loop.

A comparison of the convergence histories of the above


Fig. 7. Shape Optimization of an Isolated Airfoil: The front of nondominated
EA variants can be found in fig. 10. It is clear that, using the
solutions computed by the M(K)AEA(K). Kernel PCA to assist both metamodels and evolution operators,
enables the optimization algorithm to perform better, that is
to find better solutions with the same computational budget.

0.902
EA
0.9 EA(k)
MAEA
M(k)AEA(k)
CD/CD,ref 0.898

0.896

0.894

0.892

0.89
40 80 120 160 200
CFD Evaluations
Fig. 8. Shape Optimization of an Isolated Airfoil: Isopressure coefficient
areas around the airfoil with Cl = 1.082, Cd = 0.0845 from the front of
nondominated solution of fig. 7. Fig. 10. Shape Optimization of the fastback DrivAer Car: Comparison of the
convergence history of the optimizations performed in terms of the number
of CFD evaluations which is practically proportional to the CPU cost.

were allowed to vary along all three Cartesian directions. A comparison between the pressure fields on the optimal
All other control points were kept fixed, in order to ensure and baseline shapes of the car is shown in fig. 11.
smooth transition between deformable and nondeformable
areas, leading to 81 design variables in total. Due to symmetry, VI. C ONCLUSION
only half of the car is modeled. Each candidate solution was Two methods capable of improving the efficiency of EAs
evaluated with the incompressible fluid flow variant of the and MAEAs, when applied to high-dimensional (engineering)
same GPUenabled NavierStokes solver [16]; a steady flow optimization problems, were demonstrated. They rely on the
solver was used; depending on the geometry of the rear part of KPCA, which is repeated in each generation using the current
the car, mild flow unsteadiness might appear, in which case the offspring population. The KPCA can (a) assist the evolution
timeaveraged drag of the last iterations is used as objective operators by mapping the populations to a feature space,
function. The SpalartAllmaras [18] turbulence model is used. in which the problem appears to be more separable and
The computational mesh consists of 1.4106 nodes and each (b) reduce the number of design variables with which the
evaluation takes 40min on an NVIDIA K20 GPU, including metamodels are trained. The Kernel PCA has the ability to
the morphing process. process the design variables more efficiently than the Linear
The optimization was carried out using a (10, 20) EA, PCA, thus, the evolution operators perform better in the
EA(K), MAEA, MAEA(K) and M(K)AEA(K). The stopping KPCAs feature space. This can lead to a better performing
criterion was 200 evaluations on the CFD model. In the MAEA EAbased optimization, as shown in the presented problems.
IP E
variants, TDB = 20 and e = 3. In the runs supported by the In MAEAs, the number of design variables used to train the
P CA
PCA, gP CA = 2 and, when metamodels are used, TDB = 20, metamodels can be significantly reduced, leading to more
NDR = 71. accurate predictions. The latter can be seen by comparing
R EFERENCES
[1] M.K. Karakasis, A.P. Giotis and K.C. Giannakoglou. Inexact information
aided, low-cost, distributed genetic algorithms for aerodynamic shape
optimization. International Journal for Numerical Methods in Fluids,
43(10-11):11491166, 2003.
[2] Y. Jin, M. Olhofer and B. Sendhoff. A framework for evolutionary
optimization with approximate fitness functions. IEEE Transactions on
Evolutionary Computation, 6(5):481494, 2002.
[3] M. Emmerich, K.C. Giannakoglou and B. Naujoks. Single- and
multi-objective evolutionary optimization assisted by Gaussian random
field metamodels. IEEE Transactions on Evolutionary Computation,
10(4):421439, 2006.
[4] S. Haykin. Neural Networks: A Comprehensive Foundation. Prentice
Hall, New Jersey, USA, 2nd edition, 1999.
[5] M.K. Karakasis and K.C. Giannakoglou. On the use of metamodel-
assisted, multi-objective evolutionary algorithms. Engineering Optimiza-
tion, 38(8):941957, 2006.
[6] R.H. Myers and D.C. Montgomery. Response Surface Methodology:
Process and Product Optimization Using Designed Experiments. 2nd
Fig. 11. Shape Optimization of the fastback DrivAer Car: Comparison of the Ed. New York, 2002.
pressure field on the optimal (right) and the baseline (left) shape of the car. [7] S.A. Kyriacou, S. Weissenberger and K. C. Giannakoglou. Design of a
The rear part of the car exhibits higher pressure and this reduces the drag. Matrix Hydraulic Turbine using a Metamodel-Assisted EvolutionaryAl-
gorithm with PCA-DrivenEvolution Operators. International Journal of
Mathematical Modelling and Numerical Optimization 3 (2): 45-63.
[8] K. Deb, A. Pratap, S. Agarwal and T. Meyarivan. A fast and elitist multi-
the metamodels relative prediction error between MAEA objective genetic algorithm NSGA-II. IEEE Transactions on Evolutionary
and M(K)AEA; see for instance fig. 12 for the Weierstrass Computation, 6(2):182-197, 2002.
[9] E. Zitzler, M. Laumans and L. Thiele. SPEA2: Improving the Strength
problem of section IV. The EA variants supported by the Pareto Evolutionary Algorithm for Multiobjective Optimization. EURO-
KPCA outperform not only the conventional EAs but those GEN 2001. Athens, Greece, September 19-21, 2001.
assisted by the Linear PCA too, as it becomes clear from the [10] J. Branke and C. Schmidt. Faster convergence by means of fitness
estimation. Soft Computing, 9(1):1320, 2005.
results of the mathematical optimization problems. In real [11] S.A. Kyriacou and V.G. Asouti and K.C. Giannakoglou. Efficient PCA-
world applications, such as the optimization of aerodynamic driven EAs and metamodel-assisted EAs, with applications in turboma-
shapes, usually with a great number of design variables and chinery. Engineering Optimization 46 (7): 895-911, 2013.
[12] B. Scholkopf, A. Smola and K. Muller. Nonlinear Component Analysis
costly evaluations, the EA or MAEA which make use of the as a Kernel Eigenvalue Problem. Technical Report, Max-Planck-Institute
KPCA can be proved absolutely necessary for these problems fur biologische Kybernetik, 1996.
to be solved at affordable computational cost. Note that the [13] Q. Wang. Kernel Principal Component Analysis and its Applications in
Face Recognition and Active Shape Models. CoRR, 1207(3538), 2012.
computational cost for the KPCA is negligible compared to [14] The EASY (Evolutionary Algorithms SYstem) software,
the cost of a single evaluation, in this kind of problems which http://velos0.ltt.mech.ntua.gr/EASY.
involve computational expensive CFD software. [15] P. Joshi, M. Meyer, T. DeRose, B. Green and T. Sanocki. Harmonic
Coordinates for Character Articulation. 34th International Conference
and Exhibition on Computer Graphics and Interactive Techniques, San
Diego, California USA, August 7-9, 2007.
[16] I.C. Kampolis, X.S. Trompoukis, V.G. Asouti and K.C. Giannakoglou
CFD-based analysis and two-level aerodynamic optimization on graphics
M(K)AEA processing units. Computational Methods Applied Mechanical Engineer-
MAEA ing, 199(9-12):712-722, 2010.
1000 [17] K.T. Tsiakas, X.S. Trompoukis, V.G. Asouti and K.C. Giannakoglou.
Relative error (%)

Shape Optimization of Wind Turbine Blades using the Continuous Adjoint


Method and Volumetric NURBS on a GPU Cluster. EUROGEN 2015,
Glasgow, UK, September 14-16, 2015.
100 [18] P. Spalart and S. Allmaras. A one-equation turbulence model for
aerodynamic flows. AIAA Paper 92-0439, 1992.

10

1
0 50 100 150 200 250 300
Generations

Fig. 12. Comparison of the relative prediction error (averaged over the entire
offspring population in each generation) of the metamodels with and without
the use of the KPCA. Computed by processing the MAEA and M(K)AEA
results of the Weierstrass problem.

You might also like