You are on page 1of 14

Computers and Chemical Engineering 126 (2019) 204–217

Contents lists available at ScienceDirect

Computers and Chemical Engineering


journal homepage: www.elsevier.com/locate/compchemeng

Surrogate modeling of phase equilibrium calculations using adaptive


sampling
Corina Nentwich∗, Sebastian Engell
Department of Biochemical and Chemical Engineering, TU Dortmund University, Emil-Figge-Str. 70, Dortmund 44221, Germany

a r t i c l e i n f o a b s t r a c t

Article history: Equation of state models as the Perturbed-Chain Statistical Associating Fluid Theory (PC-SAFT) model are
Received 1 December 2018 accurate and reliable prediction models for phase equilibria. But due to their iterative nature, they are
Revised 4 April 2019
difficult to apply in chemical process optimization, because of long computation times. To overcome this
Accepted 8 April 2019
issue, surrogate modeling – replacing a complex model by a black-box model – can be used. A novel sur-
Available online 11 April 2019
rogate modeling strategy for phase equilibria is presented, combining the training of a classifier model
Keywords: with regression models for the phase composition using a mixed adaptive sampling method. We discuss
Surrogate models the selection of the parameters of the sampling algorithm and a suitable stop criterion for the exam-
Adaptive sampling ple ternary liquid-liquid equilibrium system of n-decane, dimethylformamide and 1-dodecene in detail.
Phase equilibria The sequential mixed adaptive sampling method is compared to the one-shot Latin hypercube sampling
PC-SAFT design.
© 2019 Elsevier Ltd. All rights reserved.

1. Introduction models, the density root problem as well as the phase equilib-
rium conditions must be fulfilled, which requires the use of em-
In computer-based process optimization, the reliability of the bedded calculations that lead to a significant computational effort.
optimization result depends on the quality of the process model. In This makes these advanced thermodynamic models difficult to use
order to obtain an accurate representation of the process, models for process optimization.
based on first principles are usually preferred. In order to overcome this issue, the surrogate modeling
In the modeling of chemical processes, phase equilibria play an methodology can be applied. Surrogate modeling is understood
important role. For example, the solubility of a feed material in here as replacing a complex model by a simpler black-box model.
the reaction solution significantly influences the speed of reaction, One can distinguish between two different classes of surrogate
and the accurate computation of the composition of the vapor and modeling problems: classification problems with two or more dis-
liquid phases in equilibrium is fundamental to the modeling of dis- crete outputs and regression problems with one or more continu-
tillation columns. ous outputs that are approximated.
For phase equilibrium calculations, activity coefficient models As the quality of the surrogate model is dependent on the
or equations of state models can be employed. Activity coefficient choice of the training points, sampling is an important aspect of
models require less computational effort, but are not applicable at the surrogate modeling process. Since sampling involves computa-
elevated pressures, close to critical temperatures, and for multi- tionally intense evaluations of the original function, the sampling
component systems. For such systems, equations of state models objective is to sample as few points as possible with a maximum
should be preferred (Merchan and Wozny, 2016; Schäfer et al., gain of information on the modeled phenomenon.
2014). For complex phase systems, advanced equations of state There are mainly two approaches for sampling: sampling once
models as the PC-SAFT model are suitable for accurate predictions (one-shot) or adaptive sampling. In many applications, one-shot
over a broad range of operating conditions. The PC-SAFT model space-filling designs, such as the Latin hypercube sampling (LHS),
has been applied to a wide range of different systems (Kleiner Monte-Carlo or Halton sequences are used to fit surrogate mod-
et al., 2009; Kontogeorgis and Folas, 2010; Tumakaka et al., 2005). els. However, in many applications, the modeled quantities exhibit
However, in order to solve phase equilibria using equation of state complex responses as discontinuities or a strong curvature in spe-
cific regions of the input space. In space-filling designs, such com-
plex structures are often not approximated well. For these cases,

Corresponding author. adaptive or sequential sampling designs can be used. In this con-
E-mail address: corina.nentwich@tu-dortmund.de (C. Nentwich). text, the terms of exploration – sampling in a space-filling man-

https://doi.org/10.1016/j.compchemeng.2019.04.006
0098-1354/© 2019 Elsevier Ltd. All rights reserved.
C. Nentwich and S. Engell / Computers and Chemical Engineering 126 (2019) 204–217 205

ner – and exploitation – more dense sampling in regions with j


where xi is the molar fraction of component i = 1, . . . , NC in phase
complex behavior of the original function – are used. The trade- j (feed, phase A or phase B),  is the phase fraction of phase A.
off between exploration and exploitation is a recent field of re- On the other hand, the thermodynamic equilibrium conditions
search. Cozad et al. (2014) and Kleijnen and Van Beers (2004) fo- (Eq. (3)–(5)) have to hold.
cus on exploitation. The mixed adaptive sampling approach pro-
posed in Eason and Cremaschi (2014), which was further improved TA = TB (3)
by Jin et al. (2016), uses two equally weighted sampling objectives,
while the recent work of Garud et al. (2017) proposes an approach pA = pB (4)
of adaptive weighting of the two sampling objectives.
Prior to sampling, the input and output variables of the sur- μAi = μBi , (5)
rogate modeling problem have to be defined. For a chemical sys-
where Tj ,andpj μij
are the temperature, pressure and chemical
tem, the composition, temperature and pressure define the num-
potential of component i in phase j. Inserting the definition of the
ber of phases and the composition of these. Modeling the phase
chemical potential μi as given in Eq. (6), Eq. (5) leads to the isofu-
j
composition of a mixture in the range where the transition be-
gacity criterion in Eq. (7):
tween a one-phase and a two-phase mixture occurs with a regres-
 
sion model leads to a discontinuity in the output space. In general, f ij
discontinuities in the output space are hard to approximate with μ = μ (T , p ) + RT · ln
i
j id
i
re f
(6)
pre f
surrogate models in an accurate fashion. Restricting the valid input
range to a region which is biphasic in any case as in other work fiA = fiB , (7)
(Nentwich and Engell, 2016) avoids discontinuities in the output where μid
i
is the ideal chemical potential at temperature T and ref-
space, but overconstrains the operating range which is not desir- erence pressure pref , R is the universal gas constant and fi is the
j
able in process optimization. fugacity of component i in phase j.
To cope with this problem, a novel surrogate modeling strategy In order to solve the isofugacity criterion (Eq. (7)), the fugacity
for phase equilibria is investigated in this contribution. A classifier for each component i has to be expressed for each phase. Since
is trained on the obtained data in order to identify the biphasic this is specific for the chosen thermodynamic model, it is shown
region. In a second step, regression models are used to model the in more detail for equation of state models in the next subsection.
compositions of the two phases if the point is classified as being
biphasic. This approach makes the use of equation of state mod- 2.2. Application of equation of state models
els possible within chemical process simulation and optimization,
since the computationally expensive frequent solution of the phase j
The fugacity fi of a component i in phase j can be expressed in
equilibrium is avoided. To obtain accurate classifier and regression
terms of its fugacity coefficient ϕi as shown in Eq. (8).
j
models without calling the original thermodynamic model exten-
sively for sampling, a mixed adaptive sampling scheme is applied. fij = xij ϕij p j . (8)
As a case study, the process of the hydroformylation of 1-
dodecene performed in a thermomorphic solvent system is exam- Since the following equations in this section are all phase-specific,
ined. The proposed surrogate modeling strategy is demonstrated the index for the phase j is not shown for convenience.
for the modeling of the liquid-liquid equilibrium of the ternary sys- The thermodynamic behavior of a system can be described by
tem n-decane, dimethylformamide and 1-dodecene, which is cal- modeling the Helmholtz energy. The fugacity coefficient ϕ i can also
culated using the PC-SAFT equation of state model. For the consid- be formulated in terms of the residual Helmholtz energy ares as
ered case study, the mean computation times were 5.1 seconds per shown in Eq. (9).
phase equilibrium calculation using PC-SAFT which could be re-     
∂ ares 
NC
∂ ares
duced to 0.003 seconds by applying the developed surrogate mod- ln ϕi = ares + − xj + Z − 1 − ln Z, (9)
els on a standard computer (Windows 7, 3.6 GHz dual core In- ∂ xi j=1
∂xj
tel(R) i7, 18 GB RAM).
where xi is the molar fraction of component i in the phase and Z
is the compressibility factor of the phase.
2. First principles modeling of phase equilibria The compressibility factor Z can be calculated by solving the
density root problem:
In this section, the phase equilibrium problem is defined and  
the solution procedure when applying equation of state models, ∂ ares
Z =1+ρ (10)
e. g. PC-SAFT, is described. The results of this iterative procedure ∂ρ
are used to provide data to train a surrogate model which com-
p = Zkb T ρ , (11)
putes the equilibrium by a simple function call.
where ρ is the molar density number, kb is the Boltzmann con-
stant, p is the pressure and T is the temperature.
2.1. The phase equilibrium conditions
Depending on the chemical system considered, a thermody-
namic model for the Helmholtz energy ares must be chosen.
The simplest approach to describe a phase equilibrium is a
Here we use the PC-SAFT equation of state model, since the
standard two-phase flash model in which the phase equilibrium
model is capable of not only modeling temperature-dependent LLE,
problem can be divided into two sub problems. On the one hand,
but also the solubility of supercritical components such as the syn-
the component and mass balance conditions (Eq. (1)–(2)) have to
thesis gas in the reaction mixture of the considered process de-
be fulfilled.
scribed in Section 5.
i = xi + (1 − )xi
xin A B
(1)
2.2.1. PC-SAFT
NC 
  The Perturbed-Chain Statistical Associating Fluid Theory
0= xAi − xi ,
B
(2) (PC-SAFT) equation of state model is one of the most detailed
i=1 predictive models for phase equilibrium calculations. The model
206 C. Nentwich and S. Engell / Computers and Chemical Engineering 126 (2019) 204–217

Fig. 1. Scheme of the solution procedure when applying a flash model based on an equation of state model.

was developed by Gross and Sadowski (2001). In this approach, j


xi by using Eqs. (1) and (13). For this new composition, the den-
a molecule is modeled as a chain which is formed by spherical sity root problem is again solved. This is repeated until the values
segments. The interaction of different chains is described by j
xi do not change significantly anymore.
the repulsive behavior of a hard-chain reference system in the j
The new values for xi are taken as new values for solving the
perturbation theory of Barker and Henderson (1967). Besides
overall balance equations in the next iteration of the process solver,
this repulsive hard-chain interaction (hc), the attractive inter-
which will have an effect on the values of all process variables in
action term of the dispersion (disp) is considered and a series
this iteration. If they change, the thermodynamic model is called
of additional attractive interactions can be accounted for, e. g.
again. The overall process model is solved as soon as the chosen
association (assoc) and polar interactions (dipole). j
values for xi at the chosen process conditions are equal to the so-
The residual Helmholtz energy (ares ) is described as the sum of
lution of the flash model call at these conditions. As this takes sev-
the different interactive effects, as shown in Eq. (12):
eral iterations of the process solver and thus several calls of the
ares = ahc + adisp + aassoc + adipole . (12) thermodynamic model, this procedure is time-consuming for pro-
cess models using few phase equilibria calculated with an equa-
tion of state and prohibitively slow in the optimization of complex
2.2.2. The solution procedure using PC-SAFT
process models with many phase equilibria. Therefore it is advan-
The solution of the phase equilibrium is done by an external
tageous to replace this implicit phase equilibrium calculation by an
model call of a standard two-phase pT-flash model implementation
explicit surrogate model.
of Merchan and Wozny (2016), using the MATLAB toolbox AdiMat
(Bischof et al., 2002) for the partial derivatives of ares in Eqs. (9)–
2.3. The local model methodology
(10).
The equation system given by Eqs. (1), (2) and (7) can be re-
There are different possibilities of introducing explicit approx-
duced by introducing equilibrium factors κ ,
imation models in order to reduce the computational effort that
ϕiA xAi is needed to solve the phase equilibrium problem. The local
κi = = . (13) model methodology was originally introduced to provide an ap-
ϕiB xBi
proximation of κ i (Eq. (13)). The first approaches used simplified
The resulting equation system is called the Rachford-Rice equation concentration-independent expressions for κ i (Boston and Britt,
(Eq. (14)) which has to be solved for the phase fraction of phase A 1978; Leesley and Heyen, 1977). Since these models have been
 (see also Eq. (2)), shown to be not very accurate, they were extended by e.g. modify-
ing the Margules model to introduce concentration-dependencies

NC
xin κ − 1)
i ( i (Chimowitz et al., 1983; 1984). Since these models are in general
= 0. (14)
1 −  + κi very limited with respect to their range of validity, e.g. because
i=1
of being related to a reference state, the model parameters have
The iterative calculations that are necessary to simulate a process to be updated as soon as the solution procedure has moved away
model using a flash model call to solve the phase equilibrium are from the current range of validity. In later work, the expression lo-
summarized in Fig. 1. cal model is also used as a term for linearized transformations or
In order to solve the overall balance equations, the process simplified partial derivatives of thermodynamic properties in gen-
solver chooses values for all process variables in the 1st itera- eral. Several works estimate the optimal update method and fre-
tion. Given the process conditions, the thermodynamic model is quency, e.g. (Macchietto et al., 1986; Støren and Hertzberg, 1994).
called. For dynamic simluation and control studies, the application of such
First, the density root problem (Eqs. (10)–(11)) is solved to ob- simple models is advantageous, since the states do not change
tain values for the compressibility factor Z and the density number very fast during the solution procedure. For process design and
ρ of each phase. Then it is possible to calculate the fugacity coeffi- optimization, this is not feasible due to the fast change of states
cients ϕi for each component i in phase j with the given composi-
j
(Perregaard et al., 1992).
tion xi , the compressibility factor Zj and density number ρ j at the
j
Using approximations which are valid in a broader range to
conditions T and p (see Eq. (9)). The equilibrium factors κ i are cal- avoid parameter update makes it necessary to employ more com-
culated (Eq. (13)) and the Rachford-Rice equation (14) is solved for plex model structures. Also, the phase equilibrium still has to be
the phase fraction . This value is used to update the composition solved during the solution of the balance equations of the process.
C. Nentwich and S. Engell / Computers and Chemical Engineering 126 (2019) 204–217 207

Our surrogate modeling approach is a promising alternative and is sampling frequency might be sufficient, the behavior of the orig-
discussed in detail in the next section. inal function at larger values of x is not captured by the few sam-
ples in this region, the local minima and the local maximum are
2.4. The surrogate modeling problem missed. The sampling objective to cover such regions with a more
dense sampling design is referred to as exploitative sampling de-
First a decision on the inputs and outputs of the surrogate sign. Since knowledge about the behavior of the original function
model has to be made. The most intuitive method for representing in different regions of the input space is needed, these designs can-
phase equilibrium calculations is the approximation of the misci- not be executed in a one-shot manner. They are implemented in
bility gap. Analogous to the last section, the number of outputs sequential sampling approaches which are reviewed in Section 3.2.
can be reduced by introducing the distribution coefficient of each Since both exploratory and exploitative sampling objectives are im-
component Ki (Eq. (15)) as the output instead of the molar fraction portant to achieve a good performance of the surrogate model with
xj,i of the component i in each phase j: few sampling points, they are combined in this work. Our algo-
rithm is based on the approach of Eason and Cremaschi (2014),
n˙ i,A xi,A · n˙ A xi,A xi,B − xi,in which is explained in more detail in Section 3.3.1. Modifications of
Ki = = = · , (15)
n˙ i,in xi,in · n˙ in xi,in xi,B − xi,A the algorithm are explained in Section 3.3.2 and the stop criterion
where n˙ i,A and n˙ i,in are the molar flows of component i in phase is discussed in Section 3.3.3.
A and in the feed of a continuous separation process. n˙ A and n˙ in
denote the total molar flows of phase A and of the feed. 3.1. One-shot designs
The distribution coefficient is only defined within the miscibil-
ity gap, which is changing its shape for different temperatures. In Since agglomerations of samples can occur in random sampling
order to model the shift of the biphasic region, a classifier is intro- designs as Monte Carlo methods, most one-shot sampling designs
duced. The distribution coefficient can then be approximated by a uniformly cover the input space. They are based on geometric pat-
continuous regression model and the regression model is used only terns as dividing the inputs space into grids or on the use of math-
if the classifier assigns the point to the biphasic region, otherwise ematical characteristics of sampling functions.
the concentrations remain unchanged. The latter is applied in quasi-random sampling methods which
Prior to the training of the models, sample points have to be share the characteristic of a low discrepancy. The discrepancy of a
selected and the original thermodynamic model must be evaluated sequence is connected to the uniform distribution of the sample
for these points which is time-consuming. Approaches to the selec- locations, as it tends to zero when the number of samples tends to
tion of sample points are discussed in the next section. infinity in a fixed interval (Niederreiter, 1992).
One example is the Halton sequence sampling method
3. Sampling (Halton, 1960) which is commonly applied in surrogate modeling
problems. The Halton sequence sampling is choosing the ith sam-
Especially for the case of expensive model evaluations, the goal ple location in the input space, xi , based on the Halton sequence
of devising a sampling procedure is to sample as few points as
 
possible but nonetheless get a model of good accuracy over the xi =  p1 (i ),  p2 (i ), . . . ,  pd (i ) , (16)
full range of inputs of interest. As the training locations have a
strong effect on the accuracy of the surrogate model, finding the where pj is the jth radical inverse function (see Eq. (17)) with
best sampling locations is an important aspect. A common ap- pj as base, being the jth prime number and d is the input space
proach is to equidistantly cover the complete input-space which dimension.
is referred to as space-filling or exploratory sampling design. This

inf
approach has the advantage that the original model can be eval-  p j (i ) = ak (i ) p−k
j
−1
, (17)
uated in a one-shot manner, since no previous knowledge of the k=0
original function is needed. A brief overview of one-shot space-
filling sampling designs is given in Section 3.1. The drawback of where ak (i) are the digit expansion coefficients of the integer i in
this sampling method is that the behavior of the original func- base pj ,
tion is not taken into account in the selection. The consequence

inf
is that regions of complex structures as discontinuities or a strong i= ak (i ) pkj . (18)
curvature are often not covered by enough samples to capture the k=0
true behavior with the surrogate model. An illustrative example is
shown in Fig. 2. This sampling design is deterministic and the number of samples
In this example, the 1-dimensional input x is mapped to an 1- can be chosen freely.
dimensional output y. While for smaller values of x the chosen Full-factorial or partial-factorial sampling designs address the
space-filling objective by dividing the input space into discrete lev-
els and locate samples at combinations of these different levels.
This provides a good space coverage, but at the cost of having a
huge sample set size, being the number of levels to the power of
the number of input features in the case of a full factorial design.
In Latin hypercube sampling (LHS), the input space is also di-
vided into discrete levels. However, in contrast to classical factorial
designs, the LHS design does not use all combinations of these lev-
els. Instead, the intervals for each factor in the input space are de-
fined as layers. In each layer only one sample is placed randomly.
Fig. 3 shows two examples of LHS designs for a 2-dimensional in-
put space.
Fig. 2. Illustrative example of a space-filling sampling design (cross: sample loca- As a design as shown in Fig. 3b should be preferred over a de-
tion). sign as shown in Fig. 3a, an optimization is performed, maximizing
208 C. Nentwich and S. Engell / Computers and Chemical Engineering 126 (2019) 204–217

3.3. Mixed adaptive sequential design

The applied mixed adaptive sequential sampling approach is


based on an algorithm (Eason and Cremaschi, 2014) which com-
bines the space-filling objective with an estimation of the model
prediction variance for choosing new points.

3.3.1. Principle
In each iteration, c new candidates are proposed without eval-
uating the original function. For each candidate j, j = 1, . . . , c, the
minimum distance dj to the n = 1, . . . , N points that are already
Fig. 3. LHS design examples for a 2-dimensional input space. present in the current design and the jackknife variance s2j of the
model predictions for each candidate are determined. The ith sub-
set model, i = 1, . . . , NSS, is trained on NSS − 1 subsets, analogous
the minimum distance between sample points. Thus, the space-
to cross validation methods, e. g. subset model 1 is trained on sub-
filling objective is reached and the number of samples can be cho-
sets 2 to NSS, leaving out the samples in subset 1 for training.
sen freely. A review of one-shot sampling designs can be found in
Candidates with the largest value of ηj (see Eq. (20)) are selected,
Simpson et al. (1997). To combine the exploratory sampling objec-
the corresponding outputs are calculated and the procedure is re-
tive with an exploitative sampling objective, the sampling proce-
peated until a stopping criterion is met.
dure has to be done in a sequential manner. Sequential sampling
designs are discussed in the next section. dj s2j
ηj = + , (20)
max j d j max j s2j

3.2. Sequential designs where dj is the minimum Euclidean distance (Eq. (21)) between
the candidate and the current design point and s2j is the jackknife
There are applications in which it is reasonable to use purely variance (Eq. (22)) calculated by the weighted average (Eq. (23))
space-filling sequential sampling designs (see Crombecq et al., of the jackknife pseudo values y˜ ji of each candidate j predicted by
2011), but generally they are addressing the trade-off between ex- subset model i (Eq. (24)).
ploration and exploitation. Starting from a small initial sampling
design, new sample points are chosen from a set of candidates for dj = min x j − xn 2 (21)
n∈{1,...,N}
which the true function evaluation has not yet been performed.
NSS 
 2
Since the behavior of the original function is unknown prior to 1
s2j = · y˜ ji − y˜¯ j (22)
the function evaluations and the initial sampling design consists NSS(NSS − 1 )
i=1
only of few points, complex regions have to be identified by an
1 
NSS
exploratory part of the sampling objective before they can be ex-
y˜¯ j = · y˜ ji (23)
ploited by an exploitative part of the sampling objective. NSS
i=1
The nature of the sampling objective is in many works de-
termined based upon characteristics of the chosen surrogate y˜ ji = NSS · yˆ(0) − (NSS − 1 ) · yˆ(−i ) , (24)
model. where yˆ(0 ) is the prediction for candidate j of the overall model,
For example, the expected improvement function (Eq. (19)) that trained on all i = 1, . . . , NSS subsets and yˆ(−i ) is the prediction for
was introduced as a sampling objective by Jones et al. (1998) uses candidate j of subset model i.
the ability of Kriging surrogates (see Section 4) to give an estimate The candidates with the highest value of ηj are selected and
on the standard deviation of the predicted value. added to the sampling design. For these selected sampling loca-
    tions the original function is evaluated and the sampling design is
  ymin − yˆ ymin − yˆ
EI (x ) = ymin − yˆ(x ) ·  + s(x ) , (19) updated. The number of added points is determined by the selec-
s (x ) s tion factor SF (Eq. (25)).

where ymin is the current minimum value in the original function cadd = SF · N, (25)
evaluations, yˆ(x ) is the surrogate model prediction at sampling lo-
where cadd is the number of selected points and N is the number of
cation x and s(x) is the standard error predicted by the Kriging sur-
candidates. So the number of additional samples increases in each
rogate model at x.
round.
There are also methods which are in principle applicable
Modifications of this sampling method are explained in the next
for different surrogate models. Crombecq et al. (2009) use a
section.
Voronoi-mosaic for the input space in order to explore regions
with few points and connect this with local linear models at
each candidate point in order to meet the exploitative objective. 3.3.2. Extension of the algorithm
Cozad et al. (2014) apply the approach of maximizing the error In the application at hand, the classifier is trained on all sam-
between surrogate model predictions and the true function value, ples in the current design, while the regression models are trained
which is resulting in a high number of original function evalua- only on the samples in the biphasic region to avoid discontinuities
tions. in the output space. In consequence, the subsets are not of equal
In contrast, the mixed adaptive sequential sampling design that size. In order to consider this for the jackknife variance calcula-
is further developed in this work avoids this by using a predic- tion, Eqs. (22)–(24) are changed to the modified jackknife variance
tion variance estimator for the exploitative part of the sampling calculation:
objective. In the following sections, the algorithm from Eason and 
NSS
1  2
Cremaschi (2014) and the modifications proposed in this work are s2j = · y˜ ji − y˜¯ j (26)
N (N − Ni )
discussed. i=1
C. Nentwich and S. Engell / Computers and Chemical Engineering 126 (2019) 204–217 209


NSS a Kriging-based approach for the deterministic global optimization
N − Ni
y˜¯ j = NSS · y˜ ji (27) of processes including a distillation column. Cozad et al. (2014) use
i=1 I=1 N − NI the model building tool ALAMO and combine different basis func-
y˜ ji = N · y ˆ (0 ) − (N − Ni ) · yˆ(−i ) , (28) tions to model a solid sorbent adsorber in the optimization of
a carbon capture process. Beykal et al. (2018) use the optimiza-
where N is the total number of samples in the current design and
tion framework ARGONAUT (AlgoRithms for Global Optimization
Ni is the number of samples in subset i. Thus the subset model
ofcoNstrAined grey-box compUTational problems) to apply differ-
predictions of subset models that are based on more samples are
ent types of surrogate models as grey-box constraints for compu-
accounted with a higher weight for the modified jackknife variance
tationally demanding derivative-free global optimization problems.
calculation.
Olofsson et al. (2018) use Gaussian processes in the Bayesian multi-
Since both the classifier and the regression models should be
objective optimization of a biological tissue engineering applica-
considered in the exploitative criterion, the variance term for the
tion. Recent reviews of surrogate modeling applications in model-
decision criterion ηj in Eq. (20) is chosen according to Eqs. (29)–
ing, feasibility analysis and optimization are given in Bhosekar and
(30):
Ierapetritou (2018) and McBride and Sundmacher (2019).
s2j 1 s2class, j 1 s2appr, j In this work, Support Vector Machine (SVM) models are cho-
= + (29) sen for the classification problem and ordinary Kriging models are
max j s2j 2 max j s2class, j 2 max j s2appr, j
chosen for the regression problem. The choice of the models and
their structure has been taken in a pre-study which showed that
s2appr, j 1  s2appr,i, j
NC
models of this structure are suitable model approaches for the case
= , (30)
max j s2appr, j NC
i=1
max j s2appr,i, j study. Generally, the applied sampling method does not exploit any
surrogate-dependent characteristic and can thus be applied for any
where s2class, j is the modified jackknife variance of the classifier other type of surrogate model.
predictions of candidate j and s2appr,i, j is the modified jackknife
variance of the regression model predictions of candidate j for 4.1. Classification using SVM
component i = 1, . . . , NC.
The basic idea of Support Vector Machines (SVM) is to design
3.3.3. Stop criterion a hyperplane function y(x) as given in Eq. (34) that separates the
In any sequential sampling approach a criterion is training data into two groups.
needed to terminate the sampling procedure. Eason and Cre-
maschi (2014) propose the slope ratio α as: 
S
  y (x ) = ai yt,i k(x, si ) + b, (34)
 Error slope from iteration i − 1 to i 
α =  . (31) i
max(Error slope in iterations ≤ i ) 
where ai are model parameters, k(x, si ) is the kernel function
As the number of iterations increases, the slope ratio is declining. value at input x with respect to the support vector input si with
A lower threshold value ε for α is set to terminate the algorithm. i = 1, . . . , S being the index of the S support vectors and yt,i being
Since two different types of models are involved in this work, a the corresponding output to the support vector input si and b is a
classifier and regression models, the total slope ratio α tot is defined constant model parameter. The kernel function can be chosen ar-
analogous to Eq. (29)–(30) as a weighted sum of the slope ratios of bitrarily. In this work, a polynomial of 3rd order has been selected
the different models (see Eqs. (32)–(33). in the prestudy (see Eq. (35)).
1 αclass 1 αappr  3
αtot = + (32) k(x, si ) = 1 + xT si (35)
2 max(αclass ) 2 max(αappr )
The distance between the hyperplane and the nearest training data
points (margin) is maximized by choosing the support vectors and
αappr 1  αappr,i
NC
= (33) estimating the model parameters ai and b by solving a reformula-
max(αappr ) NC max(αappr,i )
i=1 tion of the optimization problem in Eq. (36) (Bishop, 2006)
The considered error metrics that are applied to α class and α appr,i 1
 
are discussed in Section 4.3. max min yn wT
(sn ) + b , (36)
w,b ||w|| n
During the course of the mixed adaptive sampling method, the
classification and regression models are trained several times. A where w is a function of ai given in Eq. (37) and
is the kernel
brief introduction into the surrogate models that are used in this function dependent decomposition function, see Eq. (38)
work is given in the next section.

S
w= ai yt,i
(si ) (37)
4. Surrogate models
i

Classifiers and regression models are commonly used to solve k(x, x ) =


(x )T
(x ). (38)
different problems in the field of chemical engineering. Examples
By the optimization, the maximum margin when applying a cer-
of classification problems in this field are e. g. fault detection and
tain number of support vectors is achieved. For estimating the
diagnosis (Chiang et al., 2004; Onel et al., 2018) and drug de-
model parameters, the MATLAB Statistics and Machine Learning
sign (Byvatov et al., 2003). The application of surrogate models
toolbox is used (MATLAB, 2017).
for regression problems to reduce the computational effort is gain-
ing increasing popularity (see Cremaschi, 2015). Henao and Mar-
avelias (2011) use neural networks to represent process units in a 4.2. Regression using Kriging models
superstructure optimization, Caballero and Grossmann (2008) use
Kriging models to model the relationship of different variables in Kriging is an interpolation approach that originates from the
a distillation column simulation and Keßler et al. (2019) propose field of geostatistics and was introducted by Krige (1951) and
210 C. Nentwich and S. Engell / Computers and Chemical Engineering 126 (2019) 204–217

Matheron (1963). The predictor function y(x) consists of a regres- where yˆi,appr (x j ) is the surrogate model prediction for test sample
sion and a correlation term, as shown in Eq. (39): input xj and yi,appr (xj ) is the actual output value for component i
from the original model. Using these error metrics for testing the
y ( x ) = f ( x )T β + r ( x )T γ , (39)
model prediction quality for a large test set of 50 0 0 original func-
where f(x) is the vector of regression basis function and r(x) is the tion evaluations, the model quality can be evaluated in this work.
correlation function vector of input x, β and γ are model parame- For calculating the stop criterion, a measure of the model qual-
ters. The correlation function vector is defined as in Eq. (40): ity has to be defined. For the calculation of the modified jack-
knife variance (Eq. (28)), subset models yˆ(−k ) are trained on all
r ( x ) = [R ( θ , s 1 , x ) , . . . , R ( θ , s m , x ) ] ,
T
(40) subset samples except for subset k. The predictions of each sub-
where R is the correlation function which is chosen to be Gaussian set model yˆ(−k ) for the subset k can then be used for calculating a
in this work (see Eq. (41)), θ ∈ Rn is a parameter vector and si are cross-validation error. In this case, the percentage of misclassified
the m input locations used for the design of the model. samples CVE(MisC) and the mean absolute cross-validation error of
component i CVE(MAE)i are applied:

n
R(θ , a, b) = exp(−θk (|a − b| )2 ) (41) 100% 
Ntrain
 
k=1 CV E (MisC ) = CV E (MisC ) x(jk ) , (46)
2Ntrain
j=1
The model parameter β in Eq. (39) is the generalized least squares
solution regarding the training set (see Eq. (42)), γ is defined as in    (−k)  (k)   
Eq. (43). CV E (MisC ) x(jk ) = yˆclass x j − yclass x(jk )  (47)
 −1
β = FtT Rt−1 Ft FtT Rt−1 yt (42)
1 
Ntrain
 
CV E (MAE )i = CV E (MAE )i x(jk ) , (48)
Ntrain
γ= Rt−1 (yt − Ft β ), (43) j=1

      
CV E (MAE )i x(jk ) = yˆi,appr x(jk ) − yi,appr x(jk ) ,
where Ft is the basis function vector, Rt the correlation function (−k )
(49)
vector and yt is the output value. The index t indicates that these
functions are evaluated at the i = 1, . . . , m training input loca- (−k ) (k ) (−k )
where yˆclass (x j ) and yˆi,appr (x(jk) ) are the predictions of the sub-
tions si and are thus constant parameters in the prediction func-
set models trained on all subsets but subset k for sample in-
tion.
put xj belonging to subset k and for component i. yclass (x(jk ) ) and
There are different variants of the Kriging approach, depending
on the form of the regression term. In simple Kriging, the regres- yi,appr (x(jk ) ) are the actual output values from the original model
sion term is a known constant β ( f (x ) = 1), whereas in ordinary for input xj belonging to subset k and for component i.
Kriging, the regression term is a constant parameter β ( f (x ) = 1) In the following, the case study to which the mixed adaptive
estimated by generalized least squares. In universal Kriging, inde- sampling method is applied for training the SVM classifier and
pendent basis functions f(x) are used. Kriging models is described.
For estimating the model parameters, the MATLAB toolbox
DACE is used (Lophaven et al., 2002).
5. Case study
Kriging models provide a direct estimation of the prediction
variance. As the jackknife variance has shown to be a superior
As a case study, the process of the hydroformylation of 1-
method in comparison (Kleijnen and Van Beers, 2004) and to
dodecene to the main product n-tridecanal has been chosen
ensure that the proposed method is applicable to any surrogate
(Kiedorf et al., 2014). This process has been developed up to the
model type, the Kriging prediction variance is not applied in this
technical realization in two miniplants in the collaborative research
work.
center/transregio 63 “Integrated chemical processes in liquid multi-
In order to have an indication of the quality of the surrogate
phase systems” InPROMPT. Two different strategies of tunable sol-
models, suitable validation measures have to be applied. The ap-
vent systems have been pursued. The reaction has been performed
plied methods are explained in the following.
in a microemulsion process by employing surfactants in the mini-
plant at TU Berlin (Illner et al., 2018; Müller et al., 2017). The
4.3. Surrogate model validation process considered here is performed in a ThermoMorphic solvent
System (TMS). The process has been modeled in detail (Hentschel
To compare the quality of different models, performance criteria et al., 2014; Kaiser et al., 2016; Kiedorf et al., 2014; McBride et al.,
are evaluated for a test set. As a quality criterion for the classifica- 2016; McBride and Sundmacher, 2015) and was realized and it-
tion models, the percentage of misclassified samples of the test set eratively optimized online in a miniplant (Dreimann et al., 2017;
Misc, as defined in Eq. (44), is used. Hernández and Engell, 2016; Zagajewski et al., 2016). It also was
investigated in several optimization studies, see e. g. (Hentschel
100% 
Ntest
Misc = |yˆclass (x j ) − yclass (x j )|, (44) et al., 2015; 2014; Keßler et al., 2017; Steimel and Engell,
2Ntest 2016).
j=1
A TMS is composed of a polar (P in Fig. 4) and a non-polar sol-
where Ntest is the test set size, yˆclass (x j ) is the classifier model pre- vent (N in Fig. 4), in this case DMF and n-decane. This leads to a
diction for test sample input xj and yclass (xj ) is the actual class of temperature-dependent miscibility gap with a middle-polar com-
test sample xj from the original model calculation. ponent (M in Fig. 4), in this case 1-dodecene and tridecanal.
For the regression models, the Mean Absolute Error MAE (as de- The reaction is performed at an elevated temperature, so
fined in Eq. (45)) with respect to the predictions of the test set is that the reaction mixture forms one homogeneous liquid reac-
used as the error criterion. tion phase, as shown in Fig. 4a. After leaving the reactor, the
1 
Ntest mixture is cooled down. Due to the temperature-dependent mis-
MAEi = |yˆi,appr (x j ) − yi,appr (x j )|, (45) cibility gap, the mixture decomposes into two liquid phases,
Ntest
j=1 as shown in Fig. 4b. The catalyst system for this process
C. Nentwich and S. Engell / Computers and Chemical Engineering 126 (2019) 204–217 211

Table 2
Considered ranges for sampling of the surrogate model input variables
temperature T, feed fractions of n-decane xn−decane,in , DMF xDMF,in and 1-
dodecene x1−dodecene,in for the LLE calculations at 1 bar.

Input variable range

T [K] 298.15 − 343.15


xn−decane,in 0−1
xDMF,in 0−1
x1−dodecene,in 0−1

Fig. 4. TMS process scheme (Brunsch, 2013).


plication of the mixed adaptive sampling method for this task are
discussed in the next sections.

6. Results

In this section, the mixed adaptive sampling algorithm is ap-


plied to the case study of the ternary LLE of 1-dodecene, n-decane
and DMF. The choice of the sampling parameters which are the
number of subsets NSS (Section 6.1.1) and the selection factor SF
Fig. 5. TMS process flow sheet (Nentwich and Engell, 2016). (Section 6.1.2) is discussed first. The obtained models are compared
with models that are based on a conventional LHS design of the
same size in Section 6.2 to see the benefits of the sequential sam-
pling approach. In order to analyze the performance of the models
(Rh(acac)(CO)2 /Biphephos) is polar, therefore the polar phase is
and to be able to compare different models, one large test set with
recycled to the reactor, while the non-polar phase, which is rich
50 0 0 sample points that have been evaluated by PC-SAFT serves
in product, is further purified in the downstream process.
as reference for the error measures MisC, the fraction of misclas-
The translation of this process into a process flow sheet is de-
sified samples (see Eq. (44)), and MAEi , the mean absolute error
picted in Fig. 5.
(see Eq. (45)). In Sections 6.1 and 6.2, the results are discussed for
In the considered process, the reactor can be operated at tem-
the mixed adaptive sampling runs that are evaluated up to a high
peratures between 358 K and 378 K and pressures between 10 bar
number of iterations. The choice of the stop-criterion is discussed
and 30 bar (Hernandez et al., 2018), while the heat exchanger and
in Section 6.3.
decanter are operated at a pressure of 1 bar and temperatures be-
tween 298.15 K and 343.15 K. The accurate description of the sol-
ubility of the supercritical reactants hydrogen and carbon monox- 6.1. Decision on sampling parameters
ide in the reaction mixture as well as the temperature-dependent
LLE are crucial for predicting the performance of the process. Dur- In order to decide on a suitable parametrization of the algo-
ing the investigation of this process, thermodynamic models accu- rithm, several mixed adaptive sampling runs were performed. The
rately describing the behavior of all phase equilibria have been de- runs for each combination of algorithm parameters SF and NSS
veloped using PC-SAFT (Schäfer et al., 2012; 2014; Vogelpohl et al., were performed each 5 times. First, the impact of the number of
2013; 2014). The goal of our work is to perform process optimiza- subsets NSS on the model performance is analyzed.
tion based upon these models. In this paper we only discuss the
replacement of the PC-SAFT LLE model of the decanter by surro- 6.1.1. Number of subsets
gate models. The number of subsets NSS is connected to the exploitative ob-
The ternary LLE of the solvents n-decane and DMF with the jective of the mixed adaptive sampling algorithm. To find a crite-
feed n-dodecene is considered here. The pure component PC-SAFT rion that reflects the impact of NSS on the selection of candidates,
parameters are shown in Table 1. the part of the variance criterion of the total criterion value η (see
The binary interaction of the components is described by the Eq. (20)) for the selected new samples cadd in each iteration pRCV
binary interaction parameters: is analyzed:

ki j = −0.0 0 0315 T /K + 0.1159 (50) s2j


pRCV = , (52)
for DMF/n-decane,
η j · max j s2j

ki j = −0.0 0 0311 T /K + 0.1128 (51) where the index j in this case refers only to selected candidates
cadd in each iteration of the algorithm. Plotting this measure in a
for DMF/1-dodecene and ki j = 0 for n-decane/1-dodecene, taken box plot over the number of iterations for a low number of subsets
from Schäfer et al. (2012). (in this case NSS = 2) in Fig. 6 reveals that the influence of the
The considered ranges for sampling of the input variables for variance criterion on the decision is changing over the course of
surrogate modeling are shown in Table 2. The results of the ap- the algorithm.

Table 1
Pure component PC-SAFT parameters of the considered components.

Component reference M (g/mol) m (−) σ (Å) ε /k (K) μ (D) T range (K)

n-decane Gross and Sadowski (2002) 142.285 4.6627 3.8384 243.87 243–617
DMF Schäfer et al. (2012) 73.095 2.3660 3.6359 312.99 4.12 300–630
1-dodecene Schäfer et al. (2012) 168.320 5.0091 3.9413 254.86 1.70 310–630
212 C. Nentwich and S. Engell / Computers and Chemical Engineering 126 (2019) 204–217

Fig. 6. Boxplot of pRCV over the iterations of the sampling algorithm for NSS = 2 Fig. 8. MisC over the iterations of the sampling algorithm for SF = 0.4 and NSS = 2
and SF = 0.4 with 25%- and 75%-percentiles (box), median (line in box), minimum and NSS = 18 showing the mean(markers), minimum and maximum(bars) of each
and maximum (bars) and outliers (markers), based on all selected points in 5 runs. 5 runs.

Fig. 9. MAEi for component i over the iterations of the sampling algorithm for
Fig. 7. Boxplot of pRCV over the iterations of the sampling algorithm for NSS = 18 SF = 0.4 and NSS = 2 and NSS = 18 showing the mean(markers), minimum and
and SF = 0.4 with 25%- and 75%-percentiles (box), median (line in box), minimum maximum(bars) of each 5 runs.
and maximum (bars) and outliers (markers), based on all selected points in 5 runs.

is discussed in the following. For analyzing the influence of the


The median is at 45 − 64% in the first 4 iterations and decreases number of subsets NSS on the surrogate model performance, we
to 0.1% in the following iterations. While for a low number of iter- distinguish between the classifier model performance and the per-
ations, the 25%- and 75%-percentiles have a broad spread of about formance of the Kriging models. In Fig. 8, the MisC with respect to
40%, they narrow to a range of  pRCV = 3% in the last iteration. the test set is shown for NSS = 2 and NSS = 18.
There are several outliers indicated, 6 out of the 380 selected sam- Differences in the performance for the runs with NSS = 2 and
ples in the 7th iteration, 20 out of the 530 selected samples in the NSS = 18 can be seen in the first 5 iterations. In the following iter-
8th iteration and 423 out of the 2040 selected samples in the last ations, the difference is minor. Regarding the spread of the perfor-
iteration. A similar behavior can be observed for a large number of mance over the each 5 runs, the spread is broader for the runs
subsets (in this case NSS = 18) as shown in Fig. 7. with the NSS = 2 configuration. At a high number of iterations,
The decrease of the median pRCV over the course of the itera- this difference decreases. For the same runs of the sampling algo-
tions is faster initially in this case than for NSS = 2. After the 4th rithm, the results of the MAEi for the three different components
iteration, the median is already at 23% and decreases to 0.1% in the are shown in Fig. 9.
last iteration. Also, a narrowing of the 25%- and 75%-percentiles Considering the mean of the results of all 5 runs, the differ-
can be seen. In the last iterations there are outliers indicated in ence between the NSS = 2 and NSS = 18 runs can mainly be seen
the box plot, 154 out of the 2040 selected samples in the last iter- in the component DMF. In iterations 2 to 4 of the mixed adaptive
ation. sampling algorithm, the performance for the predictions of KDMF is
This trend of a decreasing pRCV over the course of the algo- superior for NSS of 18. At a higher number of iterations, the dif-
rithm shows that the impact of the exploitative sampling objective ferences are minor. Similar as for the MisC, the spread of the MAEi
on the selection of points is stronger in the first iterations of the results in each iteration is broader for the runs with NSS = 2.
mixed adaptive sampling algorithm, while in later iterations only As we want to design a reliable method with as small variance
the space-filling objective is pursued by adding new sample points in the performance as possible, the value of NSS = 18 is chosen for
to the sample set. Choosing a higher NSS strengthens this effect. further analysis. The number of subsets NSS is equal to the number
The impact of NSS on the performance of the surrogate models of models that have to be trained for each subset (see Eq. (24)), so
C. Nentwich and S. Engell / Computers and Chemical Engineering 126 (2019) 204–217 213

Fig. 10. MisC over the classifier sample size of the sampling algorithm for NSS = Fig. 11. MAEi for each component i over the Kriging sample size of the sampling al-
18 and SF = 0.2 and SF = 0.6 showing the mean(markers), minimum and maxi- gorithm for NSS = 18 and SF = 0.2 and SF = 0.6 showing the mean(markers), mini-
mum(bars) of each 5 runs. mum and maximum(bars) of each 5 runs.

the computational effort of the method will increase significantly


if a surrogate model is applied which needs a computationally ex-
pensive training method which is not the case for the models used
here. In these cases, a lower value could be preferred.
In the next section, the choice of the selection factor SF is dis-
cussed.

6.1.2. Selection factor


In this section, the number of subsets is kept constant at NSS =
18 and different selection factors SF are considered. Again, the per-
formance of the classifier model and of the Kriging models are
discussed separately. Fig. 10 shows the MisC-value of the classifier
models for the mixed adaptive sampling with NSS = 18 and 5 runs
each with SF = 0.2 and SF = 0.6 respectively.
The error decreases with the sample size for both cases, but the
performance is different. The MisC value is lower for runs with a
lower selection factor of SF = 0.2 compared to the case of SF = 0.6.
Fig. 12. MisC over the classifier sample size of the sampling algorithm for
The mean MisC at a sample size of 266 is at 2.3% for SF = 0.2, NSS = 18 and SF = 0.4 and same-sized LHS design-based models, showing the
while being 3.4% for a sample size of 261 for SF = 0.6. In the last mean(markers), minimum and maximum(bars) of each 5 runs.
iterations, the MisC-difference between the two cases is minor, the
mixed adaptive approach obtains a classifier model with a mean
test set error of MisC = 1.04% at 1648 sample points with SF = 0.2 mixed adaptive sampling algorithm at a chosen number of subsets
and for SF = 0.6 MisC = 1.02% at 1712 sample points. Thus can be NSS = 18 and SF = 0.4 is compared with the conventional one-shot
concluded that a lower selection factor leads to better performing space-filling design LHS, which is discussed in the next section.
classifier models at low sample sizes.
The behavior of the Kriging models with respect to the selec-
tion factor is different. Fig. 11 shows the mean MAE-values of the 6.2. Comparision with LHS design-based models
Kriging models for the mixed adaptive sampling with NSS = 18 and
5 runs with SF = 0.2 and SF = 0.6. To compare the results of the mixed adaptive sampling algo-
In contrast to the classifier model, the Kriging models generated rithm with LHS, for the sample size in each iteration of the sequen-
at lower sample sizes are generally performing better in the case tial sampling algorithm, 5 LHS designs of the same size were set
of a higher selection factor. At a lower sample size of 104 sample up. The performance plots discussed are based on classifier mod-
points, the MAEi for component DMF is 0.012 for the SF = 0.2 runs, els and Kriging models which were trained on these LHS designs
while being at a lower value of 0.007 at a similar training sample and on the models trained by 5 runs of the mixed adaptive sam-
size of 103 sample points for the SF = 0.6 case. At higher sample pling algorithm with SF = 0.4 and NSS = 18. First considering the
sizes of more than about 240 sample points, the difference is mi- mean classifier model performance, the models trained with the
nor. mixed adaptive sampling algorithm at any sample size outperform
The results are based on the same mixed adaptive sampling the LHS design-based models as can be seen in Fig. 12.
runs as those shown in Fig. 10, but the training data sample size Especially at medium sample sizes, e. g. 521 sample points, the
is reduced since only the biphasic samples are used to train the difference in the mean performance is significant with MisC =
Kriging models. 1.1% more misclassified samples for the LHS design-based classi-
The results show that a compromise between classifier and fiers compared to the models trained by the sequential approach.
Kriging model performance has to be found. As a compromise, At high sample sizes, the difference is smaller. For training the
the medium selection factor SF = 0.4 is considered in the fur- classifier models, the mixed adaptive approach leads to a superior
ther investigation. The performance of this parametrization of the performance than using the purely space-filling design.
214 C. Nentwich and S. Engell / Computers and Chemical Engineering 126 (2019) 204–217

Fig. 13. MAEi for the Ki of the three components over the Kriging sample size of the sequential sampling algorithm for NSS = 18 and SF = 0.4 showing the mean(markers),
minimum and maximum(bars) of each 5 runs of the models trained in the mixed adaptive sampling algorithm and the models based on LHS designs.

Fig. 14. MAEi of xi of the three components in the non-polar and polar phase over the Kriging sample size of the sequential sampling algorithm for NSS = 18 and SF = 0.4
showing the mean(markers), minimum and maximum(bars) of each 5 runs of the models trained in the mixed adaptive sampling algorithm and the models based on LHS
designs.

For the Kriging models, the performance difference is not as corresponding molar fractions xi of each component i in each non-
pronounced. The comparison is shown in separate diagrams for polar and polar phase are shown in Fig. 14.
each component of the mixture in Fig. 13. In general, the results for the composition of the polar phase
It can be seen that, also for the Kriging models, the perfor- are similar for models based on both sampling approaches. For
mance using adaptive sampling is better in the medium sam- predicting the non-polar phase composition, the models obtained
ple size range around 500 samples. While in general the perfor- from the sequential approach are superior for all three compo-
mance differences for Ki -predictions are small for the components nents. In the last iteration of the mixed adaptive sampling al-
n-decane and 1-dodecene, the difference in the component DMF is gorithm, the MAEi of xn−decane in the non-polar phase is 0.0010,
larger. The fraction of biphasic samples is lower in the sampling 0.0015 for xDMF and 0.0006 for x1−dodecene . For the LHS design-
designs generated by the mixed adaptive sampling compared to based models, the MAEi are 0.0013 for xn−decane , 0.0019 for xDMF
the same-sized LHS designs. Although the training sample size in and 0.0 0 06 for x1−dodecene , although the models are based on a big-
the last iteration is lower, the performance of the Kriging mod- ger sample size of 974 in contrast to 724 in the last iteration of the
els trained within the mixed adaptive sampling design is similar mixed adaptive sampling algorithm.
to the LHS design-based models. This shows that by combining As the composition of the non-polar phase is determining the
the exploratory and the exploitative sampling objective, the perfor- design of the process design for further purification (see Fig. 5),
mance of the Kriging models is superior to LHS design-based mod- a reliable surrogate model has to be obtained. But as the quality
els at a same number of biphasic samples, e. g. the MAEn−decane is of the surrogate model is not known prior to training, the num-
at 0.0011 for a sample size of 519 in the sequential approach, while ber of samples required to obtain surrogate models with a suffi-
being at 0.0020 at a similar sample size of 490 in the LHS design. cient performance is not known when using a one-shot design. A
At the same sample sizes, the MAEi for DMF is 0.0017 in the se- comparison of the computation time of the mixed adaptive sam-
quential and 0.0041 in the LHS design-based approach, while for pling algorithm with a LHS design of the same size is shown in
1-dodecene it is 0.0010 and 0.0019. The errors with respect to the Table 3.

Table 3
Mean computation time and standard deviation of the each 5 runs when applying the mixed adaptive sampling algorithm and LHS
design for a given total sample size, performed on a standard computer (Windows 7, 3.6 GHz dual core Intel(R) i7, 18 GB RAM).

Sample size 35 49 69 97 136 190 266 372 521 729 1021 1429

Mean adaptive [min] 3.83 5.22 7.32 9.91 14.27 19.98 29.25 40.20 56.25 79.10 111.45 156.68
Mean LHS [min] 3.47 4.76 6.97 11.14 15.24 20.74 29.05 39.22 56.44 81.69 107.37 148.79
Std adaptive [min] 0.54 0.52 0.65 1.01 0.93 1.21 1.12 1.87 0.53 2.44 4.74 5.50
Std LHS [min] 0.75 0.59 1.34 1.10 1.94 1.82 2.29 3.28 2.06 14.67 8.70 5.53
C. Nentwich and S. Engell / Computers and Chemical Engineering 126 (2019) 204–217 215

Overall, the differences are not significant. As an advantage of


applying the sequential approach, the information on the current
samples in each iteration can be exploited to define a suitable stop
criterion of the algorithm. This is discussed in the next section.

6.3. Stop criterion

In the previous sections, the error measures of MisC and


MAEi were calculated regarding a large test set to ensure com-
parability of the model performance. For the practical use of
the sequential sampling method, an error estimate based on the
available sample points at the current iteration of the sequential
sampling approach should be applied. As discussed in Section 4.3,
the modified jackknife variance calculation requires the genera-
tion of NSS subset models which are each trained on NSS − 1 sub-
sets (see Section 3.3.2) and can thus be used for NSS-fold cross- Fig. 16. CVE(MAE)i (left axis) and CVE(MisC) (right axis) of the xi of the three com-
validation to calculate CVE(MAE)i and CVE(MisC) (see Eqs. (46)– ponents in the non-polar and polar phase over the iterations of run 1, the blue line
indicates a threshold value of ε = 0.01. (For interpretation of the references to color
(49)). As Eason and Cremaschi (2014) apply the slope ratio crite- in this figure legend, the reader is referred to the web version of this article.)
rion of α tot ≤ ε (see sec. 3.3.3), where ε is a lower threshold, this
stop criterion is discussed first. Since the value of ε has to be pre-
defined, a suitable value has to be found. For their case studies,
Eason and Cremaschi (2014) propose ε to be equal to 0.03, but
it has to be noted that in their work another error measure, the
summed squared error SSE, was used. To find an adequate value
for the given case, α tot is at first calculated based on the test set
error measures MAEi and MisC according to Eq. (32). The resulting
α tot for the 5 runs with the configuration SF = 0.4 and NSS = 18 is
given in Fig. 15.
The red line indicates a threshold of ε = 0.03. It can be seen
that the threshold value of 0.03 is not applicable in this case, al-
though the MAEi and MisC are not changing significantly anymore
(see also Figs. 12 and 13). Only run 2 and 3 would have stopped
at iteration 8 and 11 respectively. Running the sequential sampling
algorithm longer until this value of ε is achieved can be unpro-
ductive. The resulting surrogate model may be more accurate, but
the errors in the last iteration are in the order of magnitude of
MAEi ≈ 0.001 or lower (see Fig. 14). Measuring a liquid composi-
tion with such an accuracy is hardly possible. As a consequence, Fig. 17. LLE phase diagram of the PC-SAFT predictions (lines) and surrogate model
one could choose a higher value for ε to stop at an earlier itera- predictions (markers) at 300 K (blue) and 340 K (red). (For interpretation of the
references to color in this figure legend, the reader is referred to the web version
tion, but the user has no indication of the quality of the resulting
of this article.)
models when choosing this value and as can be seen in Fig. 15, this
leads to the different mixed adaptive sampling runs being stopped
at different iterations of the algorithm.
A more practical solution to this problem is to define an abso-
lute value for the accuracy that should be achieved. As an example,
CVE(MAE)i and CVE(MisC) for the sample set at the different itera-
tions of run 1 are shown in Fig. 16.
Defining the required accuracy to be CV E (MAE )i = 0.01 as in-
dicated by the blue line, the algorithm would be stopped in
this case at iteration 11 with CV E (MAE )n−decane = 0.005 in the
non-polar phase and 0.003 in the polar phase, CV E (MAE )DMF =
0.009 in the non-polar phase and 0.005 in the polar phase,
CV E (MAE )1−dodecene = 0.004 in the non-polar phase and 0.002 in
the polar phase and CV E (MisC ) = 0.018. By this, an estimate of the
model accuracy is available and can be used to adapt the proposed
method to case study dependent requirements.
Predictions of the phase diagram that were calculated using the
resulting surrogate models are compared to the original PC-SAFT
model predictions in Fig. 17.
The Kriging models show a good prediction accuracy for the
phase diagram at both temperatures. The classifier classified all
samples correctly in this test case.
The applied surrogate modeling strategy has shown to be suit-
Fig. 15. α tot based on the test set error measures MAEi and MisC over the iterations
of 5 runs, the red line indicates a threshold value of ε =0.03. (For interpretation able to replace expensive thermodynamic model calls. The ap-
of the references to color in this figure legend, the reader is referred to the web plication of the mixed adaptive sampling scheme leads to better
version of this article.) performing surrogate models compared to a conventional purely
216 C. Nentwich and S. Engell / Computers and Chemical Engineering 126 (2019) 204–217

space-filling design. The sequential nature of the algorithm gives sic region of the separator (see Fig. 5), while the regression models
the user more flexibility in the construction of surrogate models of predict the phase composition.
a predefined accuracy.

Acknowledgment
7. Conclusions
This work is part of the Collaborative Research Center/
Transregio 63 “Integrated Chemical Processes in Liquid Multi-
As equation of state models as PC-SAFT often are not directly
phase Systems” (subproject D1). Financial support by the Deutsche
applicable in process optimization due to the computational ex-
Forschungsgemeinschaft (DFG, German Research Foundation) is
pense of the iterative computations, this work aims at replacing
gratefully acknowledged (TRR 63).
the expensive thermodynamic model calls by explicit computations
The authors also thank Roderich Wallrath, Clemens Lindscheid,
of surrogate models. In order to combine explorative and exploita-
Maximilian Cegla, Radoslav Paulen, Marina Rantanen Modéer, Si-
tive samling objectives to find the best sample locations, the mixed
mon Wenzel, Shreya Bhatia and Anoj-Winston Gladius for their
adaptive sampling approach by Eason and Cremaschi (2014) has
support.
been extended to a novel surrogate modeling strategy for phase
equilibrium calculations which is in particular suitable for LLE
models. References
Modeling the composition of the different phases of the system
leads to discontinuities in the output space as the mixture disinte- Barker, J.A., Henderson, D., 1967. Perturbation theory and equation of state for fluids.
II. A successful theory of liquids. J. Chem. Phys. 47, 4714–4721. doi:10.1063/1.
grates at certain operating conditions. To obtain models which are 1701689.
valid in the full operating range of interest, the change in the num- Beykal, B., Boukouvala, F., Floudas, C.A., Pistikopoulos, E.N., 2018. Optimal design of
ber of phases also has to be modeled. This is done by the applica- energy systems using constrained grey-box multi-objective optimization. Com-
put. Chem. Eng. 116, 488–502. doi:10.1016/j.compchemeng.2018.02.017.
tion of a classifier model. For the prediction of the composition of Bhosekar, A., Ierapetritou, M., 2018. Advances in surrogate based modeling, feasi-
the different phases, suitable regression models are then applied. bility analysis, and optimization: a review. Comput. Chem. Eng. 108, 250–267.
The goal of our work is to train surrogate models that provide a doi:10.1016/j.compchemeng.2017.09.017.
Bischof, C., Bucker, H., Lang, B., Rasch, A., Vehreschild, A., 2002. Combining source
very high prediction accuracy over the full operating range of rel- transformation and operator overloading techniques to compute derivatives
evant process conditions so that the models can be used in pro- for MATLAB programs. In: Proceedings. Second IEEE International Workshop
cess optimization without retraining which is why nonlinear clas- on Source Code Analysis and Manipulation. IEEE, Montreal, Quebec, Canada,
pp. 65–72. doi:10.1109/scam.2002.1134106.
sification and regression models are used. This is in contrast to us-
Bishop, C.M., 2006. Pattern Recognition and Machine Learning, 1 ed. Springer, New
ing local approximations that possibly have to be retrained during York.
the course of the optimization. Of course, monitoring of the range Boston, J., Britt, H., 1978. A radically different formulation and solution of the
of the variables is still necessary in the optimization so that the single-stage flash problem. Comput. Chem. Eng. 2, 109–122. doi:10.1016/
0 098-1354(78)80 015-5.
surrogate models are not employed beyond their domain of valid- Brunsch, Y., 2013. Temperaturgesteuertes Katalysatorrecycling für die homogen
ity, where the models would have to be retrained. In this work, katalysierte Hydroformylierung langkettiger Alkene. Verlag Dr. Hut, Dissertation
this approach has been applied to the operating range of interest TU Dortmund University, Dortmund, Germany doi:10.17877/de290r-10964.
Byvatov, E., Fechner, U., Sadowski, J., Schneider, G., 2003. Comparison of support
for the ternary liquid-liquid equilibrium of n-decane, dimethylfor- vector machine and artificial neural network systems for drug/nondrug classifi-
mamide and 1-dodecene, applying a SVM classifier and ordinary cation. J. Chem. Inf. Comput. Sci. 43, 1882–1889. doi:10.1021/ci0341161.
Kriging models as the regression method, as part of the goal to Caballero, J.A., Grossmann, I.E., 2008. An algorithm for the use of surrogate mod-
els in modular flowsheet optimization. AlChE J. 54, 2633–2650. doi:10.1002/aic.
make an efficient process optimization applying a complex ther- 11579.
modynamic model as PC-SAFT for all phase equilibria in the pro- Chiang, L.H., Kotanchek, M.E., Kordon, A.K., 2004. Fault diagnosis based on fisher
cess of hydroformylation of 1-dodecene possible. discriminant analysis and support vector machines. Comput. Chem. Eng. 28 (8),
1389–1401. doi:10.1016/j.compchemeng.2003.10.002.
The choice of the parameters of the mixed adaptive sampling Chimowitz, E.H., Anderson, T.F., Macchietto, S., Stutzman, L.F., 1983. Local mod-
algorithm was discussed in detail. A large number of subsets NSS els for representing phase equilibria in multicomponent, nonideal vapor-Liquid
improves the quality of the surrogate models that are trained with and liquid-Liquid systems. 1. thermodynamic approximation functions. Ind. Eng.
Chem. Process Des.Dev. 22, 217–225. doi:10.1021/i20 0 021a0 09.
the adaptive sampling approach. For the selection factor SF a com-
Chimowitz, E.H., Macchietto, S., Anderson, T.F., Stutzman, L.F., 1984. Local models
promise between classifier and regression model performance had for representing phase equilibria in multicomponent, non-ideal vapor-liquid and
to be made. The classifier model performance is best when the SF liquid-liquid systems. 2. application to process design. Ind. Eng. Chem. Process
is small and for the regression models the opposite is true. Des. Dev. 23, 609–618. doi:10.1021/i20 0 026a034.
Cozad, A., Sahinidis, N.V., Miller, D.C., 2014. Learning surrogate models for
The models that were trained using the mixed adaptive sam- simulation-based optimization. AlChE J. 60, 2211–2227. doi:10.1002/aic.14418.
pling algorithm with the configuration NSS = 18 and SF = 0.4 were Cremaschi, S., 2015. A perspective on process synthesis: challenges and prospects.
compared to models based on conventional one-shot Latin hyper- Comput. Chem. Eng. 81, 130–137. doi:10.1016/j.compchemeng.2015.05.007.
Crombecq, K., De Tommasi, L., Gorissen, D., Dhaene, T., 2009. A novel sequen-
cube designs, revealing that the mixed adaptive approach obtains tial design strategy for global surrogate modeling. In: Proceedings of the 2009
better performing classification and regression models at a similar Winter Simulation Conference (WSC). IEEE, pp. 731–742. doi:10.1109/wsc.2009.
computational expense. 5429687.
Crombecq, K., Laermans, E., Dhaene, T., 2011. Efficient space-filling and non-
The algorithm should be stopped at an accuracy that is satis- collapsing sequential design strategies for simulation-based modeling. Eur. J.
factory for the purpose, e. g. considering the accuracy of the mea- Oper. Res. 214, 683–696. doi:10.1016/j.ejor.2011.05.032.
surements. As a stop criterion, the slope ratio turned out to be Dreimann, J.M., Hoffmann, F., Skiborowski, M., Behr, A., Vorholt, A.J., 2017. Merging
thermomorphic solvent systems and organic solvent nanofiltration for hybrid
not useful in this case. The definition of threshold values for cross-
catalyst recovery in a hydroformylation process. Ind. Eng. Chem. Res. 56, 1354–
validation errors should be preferred. 1359. doi:10.1021/acs.iecr.6b04249.
Further improvements to the algorithm could be achieved by Eason, J., Cremaschi, S., 2014. Adaptive sequential sampling for surrogate model
generation with artificial neural networks. Comput. Chem. Eng. 68, 220–232.
adjusting the weighting of the exploratory objective, the classifier
doi:10.1016/j.compchemeng.2014.05.021.
and the regression models in the overall sampling objective (see Garud, S.S., Karimi, I., Kraft, M., 2017. Smart sampling algorithm for surrogate model
Eq. (20) and Eqs. (29)–(30)). development. Comput. Chem. Eng. 96, 103–114. doi:10.1016/j.compchemeng.
The next step will be the use of the surrogate models in pro- 2016.10.006.
Gross, J., Sadowski, G., 2001. Perturbed-chain SAFT: an equation of state based on
cess optimization. The classifier model can be implemented as a a perturbation theory for chain molecules. Ind. Eng. Chem. Res. 40, 1244–1260.
constraint that defines the valid operating range within the bipha- doi:10.1021/ie0 0 03887.
C. Nentwich and S. Engell / Computers and Chemical Engineering 126 (2019) 204–217 217

Gross, J., Sadowski, G., 2002. Application of the perturbed-chain SAFT equation of Matheron, G., 1963. Principles of geostatistics. Econ. Geol. 58 (8), 1246–1266. doi:10.
state to associating systems. Ind. Eng. Chem. Res. 41, 5510–5515. doi:10.1021/ 2113/gsecongeo.58.8.1246.
ie010954d. MATLAB and Statistics and Machine Learning Toolbox Release, 2017. The Math-
Halton, J.H., 1960. On the efficiency of certain quasi-random sequences of points in Works, Inc., Natick, Massachusetts, United States.
evaluating multi-dimensional integrals. Numer. Math. 2 (1), 84–90. doi:10.1007/ McBride, K., Gaide, T., Vorholt, A., Behr, A., Sundmacher, K., 2016. Thermomor-
bf01386213. phic solvent selection for homogeneous catalyst recovery based on COSMO-RS.
Henao, C.A., Maravelias, C.T., 2011. Surrogate-based superstructure optimization Chem. Eng. Process. 99, 97–106. doi:10.1016/j.cep.2015.07.004.
framework. AlChE J. 57 (5), 1216–1232. doi:10.1002/aic.12341. McBride, K., Sundmacher, K., 2015. Data driven conceptual process design for the
Hentschel, B., Kiedorf, G., Gerlach, M., Hamel, C., Seidel-Morgenstern, A., Freund, H., hydroformylation of 1-Dodecene in a thermomorphic solvent system. Ind. Eng.
Sundmacher, K., 2015. Model-Based identification and experimental validation Chem. Res. 54, 6761–6771. doi:10.1021/acs.iecr.5b00795.
of the optimal reaction route for the hydroformylation of 1-Dodecene. Ind. Eng. McBride, K., Sundmacher, K., 2019. Overview of surrogate modeling in chemical pro-
Chem. Res. 54, 1755–1765. doi:10.1021/ie504388t. cess engineering. Chem. Ing. Tech. 91, 228–239. doi:10.10 02/cite.20180 0 091.
Hentschel, B., Peschel, A., Freund, H., 2014. Simultaneous design of the optimal re- Merchan, V.A., Wozny, G., 2016. Comparative evaluation of rigorous thermodynamic
action and process concept for multiphase systems. Chem Eng Sci 115, 69–87. models for the description of the hydroformylation of 1-Dodecene in a thermo-
doi:10.1016/j.ces.2013.09.046. morphic solvent system. Ind. Eng. Chem. Res. 55, 293–310. doi:10.1021/acs.iecr.
Hernandez, R., Dreimann, J., Vorholt, A., Behr, A., Engell, S., 2018. Iterative real-Time 5b03328.
optimization scheme for optimal operation of chemical processes under uncer- Müller, D., Illner, M., Esche, E., Pogrzeba, T., Schmidt, M., Schomäcker, R., Biegler, L.T.,
tainty: proof of concept in a miniplant. Ind. Eng. Chem. Res. 57 (26), 8750–8770. Wozny, G., Repke, J.-U., 2017. Dynamic real-time optimization under uncer-
doi:10.1021/acs.iecr.8b00615. tainty of a hydroformylation mini-plant. Comput. Chem. Eng. 106, 836–848.
Hernández, R., Engell, S., 2016. Modelling and iterative real-time optimization of a doi:10.1016/j.compchemeng.2017.01.041.
homogeneously catalyzed hydroformylation process. Comput. Aided Chem. Eng. Nentwich, C., Engell, S., 2016. Application of surrogate models for the optimization
38, 1–6. doi:10.1016/b978- 0- 444- 63428- 3.50 0 05-9. and design of chemical processes. In: 2016 International Joint Conference on
Illner, M., Schmidt, M., Pogrzeba, T., Urban, C., Esche, E., Schomäcker, R., Repke, J.-U., Neural Networks (IJCNN). IEEE, Vancouver, BC, Canada, pp. 1291–1296. doi:10.
2018. Palladium-Catalyzed methoxycarbonylation of 1-Dodecene in a two-Phase 1109/ijcnn.2016.7727346.
system: the path toward a continuous process. Ind. Eng. Chem. Res. 57, 8884– Niederreiter, H., 1992. Random number generation and quasi-Monte carlo methods.
8894. doi:10.1021/acs.iecr.8b01537. Soc. Ind. Appl. Math. doi:10.1137/1.9781611970081.
Jin, Y., Li, J., Du, W., Qian, F., 2016. Adaptive sampling for surrogate modelling with Olofsson, S., Mehrian, M., Calandra, R., Geris, L., Deisenroth, M., Misener, R., 2018.
artificial neural network and its application in an industrial cracking furnace. Bayesian multi-Objective optimisation with mixed analytical and black-Box
Can. J. Chem. Eng. 94, 262–272. doi:10.1002/cjce.22384. functions: application to tissue engineering. IEEE Trans. Biomed. Eng. 1–12.
Jones, D.R., Schonlau, M., Welch, W.J., 1998. Efficient global optimization of ex- doi:10.1109/tbme.2018.2855404.
pensive black-Box functions. J. Global Optim. 13, 455–492. doi:10.1023/a: Onel, M., Kieslich, C.A., Guzman, Y.A., Floudas, C.A., Pistikopoulos, E.N., 2018. Big
1008306431147. data approach to batch process monitoring: simultaneous fault detection and
Kaiser, N.M., Flassig, R.J., Sundmacher, K., 2016. Probabilistic reactor design in the diagnosis using nonlinear support vector machine-based feature selection. Com-
framework of elementary process functions. Comput. Chem. Eng. 94, 45–59. put. Chem. Eng. 115, 46–53. doi:10.1016/j.compchemeng.2018.10.016.
doi:10.1016/j.compchemeng.2016.06.008. Perregaard, J., Pedersen, B.S., Gani, R., 1992. Steady state and dynamic simulation of
Keßler, T., Kunde, C., McBride, K., Mertens, N., Michaels, D., Sundmacher, K., complex chemical processes. Trans. Inst. Chem. Eng. 70 (A), 99–109.
Kienle, A., 2019. Global optimization of distillation columns using explicit and Schäfer, E., Brunsch, Y., Sadowski, G., Behr, A., 2012. Hydroformylation of 1-
implicit surrogate models. Chem. Eng. Sci. 197, 235–245. doi:10.1016/J.CES.2018. dodecene in the thermomorphic solvent system dimethylformamide/decane.
12.002. phase behavior-reaction performance-catalyst recycling. Ind. Eng. Chem. Res. 51,
Keßler, T., Mertens, N., Kunde, C., Nentwich, C., Michaels, D., Engell, S., Kienle, A., 10296–10306. doi:10.1021/ie300484q.
2017. Efficient global optimization of a novel hydroformylation process. Comput. Schäfer, E., Sadowski, G., Enders, S., 2014. Calculation of complex phase equilibria of
Aided Chem. Eng. 40, 2113–2118. doi:10.1016/b978- 0- 444- 63965- 3.50354- 8. DMF/alkane systems using the PCP-SAFT equation of state. Chem. Eng. Sci. 115,
Kiedorf, G., Hoang, D.M., Müller, A., Jörke, A., Markert, J., Arellano-Garcia, H., Seidel- 49–57. doi:10.1016/j.ces.2013.04.053.
Morgenstern, A., Hamel, C., 2014. Kinetics of 1-dodecene hydroformylation in a Simpson, T.W., Peplinski, J.D., Koch, P.N., Allen, J.K., 1997. On the use of statis-
thermomorphic solvent system using a rhodium-biphephos catalyst. Chem. Eng. tics in design and the implications for deterministic computer experiments. In:
Sci. 115, 31–48. doi:10.1016/j.ces.2013.06.027. Proceedings of DETC’97 1997 ASME Design Engineering Technical Conferences.
Kleijnen, J.P., Van Beers, W.C., 2004. Application-driven sequential designs for sim- Sacramento, California, pp. 1–14.
ulation experiments: kriging metamodelling. J. Oper. Res. Soc. 55, 876–883. Steimel, J., Engell, S., 2016. Optimization-based support for process design under
doi:10.1057/palgrave.jors.2601747. uncertainty: a case study. AlChE J. 62, 3404–3419. doi:10.1002/aic.15400.
Kleiner, M., Tumakaka, F., Sadowski, G., 2009. Thermodynamic modeling of complex Støren, S., Hertzberg, T., 1994. Local thermodynamic models applied in dynamic pro-
systems. In: Lu, X., Hu, Y. (Eds.), Structure and Bonding. Springer, Berlin, Heidel- cess simulation: a simplified approach. Trans. Inst. Chem. Eng. 72 (A3), 395–401.
berg, pp. 75–108. Tumakaka, F., Gross, J., Sadowski, G., 2005. Thermodynamic modeling of complex
Kontogeorgis, G.M., Folas, G.K., 2010. Thermodynamic Models for Industrial Applica- systems using PC-SAFT. In: Fluid Phase Equilibria, 228–229, pp. 89–98. doi:10.
tion: from Classical and Advanced Mixing Rules to Association Theories, 1 ed. 1016/j.fluid.2004.09.037.
Wiley, Chichester, UK. Vogelpohl, C., Brandenbusch, C., Sadowski, G., 2013. High-pressure gas solubility in
Krige, D.G., 1951. A statistical approach to some basic mine valuation problems on multicomponent solvent systems for hydroformylation. part i: carbon monoxide
the witwatersrand. Master’s thesis, South Africa, University of Witwatersrand. solubility. J. Supercrit. Fluids 81, 23–32. doi:10.1016/j.supflu.2013.04.006.
Leesley, M., Heyen, G., 1977. The dynamic approximation method of handling vapor- Vogelpohl, C., Brandenbusch, C., Sadowski, G., 2014. High-pressure gas solubility in
liquid equilibrium data in computer calculations for chemical processes. Com- multicomponent solvent systems for hydroformylation. part II: syngas solubility.
put. Chem. Eng. 1, 103–108. doi:10.1016/0098-1354(77)80015-X. J. Supercrit. Fluids 88, 74–84. doi:10.1016/j.supflu.2014.01.017.
Lophaven, S. N., Nielsen, H. B., Søndergaard, J., 2002. DACE - A Matlab Kriging Tool- Zagajewski, M., Dreimann, J., Thönes, M., Behr, A., 2016. Rhodium catalyzed hydro-
box, Version 2.0. formylation of 1-dodecene using an advanced solvent system: towards highly
Macchietto, S., Chimowitz, E.H., Anderson, T.F., Stutzman, L.F., 1986. Local models efficient catalyst recycling. Chem. Eng. Process. 99, 115–123. doi:10.1016/j.cep.
for representing phase equilibria in multicomponent nonideal vapor-Liquid and 2015.06.014.
liquid-Liquid systems. 3. parameter estimation and update. Ind. Eng. Chem. Pro-
cess Des. Dev. 25, 674–682. doi:10.1021/i20 0 034a013.

You might also like