Professional Documents
Culture Documents
https://doi.org/10.1007/s13369-020-05270-3
Abstract
The repair method for pavements should be selected considering the structural capacity of sublayers, in addition to the con‑
ditions observed at the pavement surface, to reduce the recurrence of distress in the repaired area. However, it is practically
impossible to include the structural capacity of sublayers in the database of the pavement management system (PMS) because
this would require additional tests in all expressway sections. Therefore, an artificial neural network model for predicting the
indirect tensile strength (ITS) of the intermediate layer of all asphalt pavement sections in an expressway was developed in
this study, taking the international roughness index, rut depth, surface distress, and equivalent single axle load as independent
variables. The ITS of specimens cored from target sections was measured in the laboratory, and the PMS data for the target
sections were collected. The ITS was predicted by conducting a feedforward process prior to the training step. When the
error between the predicted and measured ITSs exceeded the allowable error, the model was repetitively trained using the
resilient backpropagation method until the error fell within the acceptable boundary. The model was validated by analyzing
the correlations between the ITSs predicted from the data of the training and test sets. Finally, the model was complemented
by the corresponding minimum and maximum values of the ITS measured at the target section.
Keywords Artificial neural networks · Indirect tensile strength · Asphalt pavement · Intermediate layer · Pavement
management system
13
Vol.:(0123456789)
Arabian Journal for Science and Engineering
13
Arabian Journal for Science and Engineering
13
Arabian Journal for Science and Engineering
Fig. 4 Distribution of variables
13
Arabian Journal for Science and Engineering
ITS (MPa) Target sections 88 2.07 1.55 1.21 0.87 0.19 1.21 0.45 − 0.07 − 0.81 Yes
IRI (m/km) Target sections 88 2.69 1.74 1.37 1.17 0.66 1.47 0.44 0.69 − 0.12 Yes
Entire sections 54,262 7.66 1.91 1.45 1.17 0.45 1.61 0.63 1.40 2.81 Yes
RD (mm) Target sections 88 12.55 7.22 5.91 4.16 3.00 6.05 2.18 0.71 0.03 Yes
Entire sections 54,262 20.00 6.48 4.82 3.57 0.00 5.32 2.46 1.38 2.93 Yes
SD (m2) Target sections 88 15.94 0.35 0.04 0.00 0.00 0.86 2.36 4.34 21.76 No
Entire sections 54,262 83.72 0.12 0.00 0.00 0.00 0.64 2.48 8.87 128.74 No
Cube root of SD Target sections 88 2.52 0.70 0.33 0.00 0.00 0.47 0.59 1.38 1.51 Yes
Entire sections 54,262 4.37 0.49 0.00 0.00 0.00 0.31 0.54 2.04 4.26 Yes
ESAL Target sections 88 7185 4393 3980 3811 2113 4136 1137 1.33 2.33 Yes
Entire sections 54,262 11,638 4667 3400 2280 28 3592 1826 0.41 − 0.33 Yes
The variables, except for SD, satisfied the criterion of nor‑ was developed using the data of the training set according
mality for both target sections and entire sections, as summa‑ to the procedure shown in Fig. 5. The structure of the model
rized in Table 2, and showed normal distributions, as shown comprises three layers: one input layer, one hidden layer, and
in Fig. 4a, b, and e. However, the skewness and kurtosis of SD one output layer, as shown in Fig. 6. The input layer com‑
were 4.26 and 21.76 for target sections and 8.70 and 124.33 prised four nodes matching the four independent variables,
for the entire sections, respectively, thereby failing to satisfy IRI, RD, SD, and ESAL, while 3 to 20 nodes were tried in
the criterion of normality. The SD distribution curve leaned the hidden layer in turn. The output layer comprised only
toward the left, as shown in Fig. 4c because SD was zero at one node for the dependent variable, ITS.
nearly all unit sections, and a significantly high SD was cal‑ The ESAL, with values in the range of thousands, had
culated at only a few unit sections owing to several distresses. a larger scale than the IRI, RD, and SD, with single-digit
The SD that did not satisfy the criterion of normality values, when comparing the x axes in Fig. 4. A problem,
could be applied in this study because the ANN method is for which the ITS was predicted mainly by ESAL, occurred
free from the distribution of variables. However, the pre‑ because ESAL influenced the predicted ITS more than the
diction accuracy of the ANN model was improved by the other independent variables because of its larger scale [25].
appropriate transformation of the variables to reduce the In addition, the SD with ordinary magnitude could be distin‑
skewness and kurtosis of the data [19–22]. Therefore, as guished as an outlier before standardization because the SD
shown in Fig. 4d, a cube root was applied to the SD based on of almost all the unit sections was zero, as shown in Fig. 4c.
the form of distribution, and the skewness and kurtosis were Therefore, standardization was conducted in order to reduce
recalculated [23]. As a result, the skewness and kurtosis of the scale of ESAL according to the robust method, as shown
the cube root of SD were 1.36 and 1.51 for target sections in Eq. (2), which used a median value that was less sensitive
and 2.00 and 4.08 for the entire sections, respectively, sat‑ to outliers than the average value [25, 26].
isfying the criterion of normality. Therefore, the cube root ( )
of SD was used as an independent variable in the model xi − median xi
Xi = ( ) (2)
instead of the SD. MAD xi
where Xi = s t a n d a r d i z e d
( ) value of x ,
| ( )|
3 Development of the ANN Model MAD = 1.4826median|xi − median xi | , median xi = median
| |
value of x.
The PMS data of the target sections, where specimens were As a result of the standardization of the independent vari‑
cored, were divided into two parts in the ratio 7:3 to classify ables with different scales, as shown in Fig. 7a, the medians
the data into a training set of 62 sections to teach the ANN of all the independent variables were set to zero, adjusting
model and a test set of 26 sections to validate the model. the scales of the independent variables, as shown in Fig. 7b.
The datasets for each group were selected by a non-restored The initial weights and biases were set as the first step
random extraction method that generates random numbers of ANN modeling. Generally, the random numbers chosen
using the Mersenne–Twister method [24]. The ANN model from the standard normal distribution curve, with an average
13
Arabian Journal for Science and Engineering
domly determined by the initial weights and biases because where i = index of input variable; 1, 2, 3, …, i, j = index of
each initial weight or bias uses a unique random number. hidden node; 1, 2, 3, …, j, o = index of output node; 1, 2,
Therefore, only one random number generated on the stand‑ 3 …, o, xi = value of input variable i, wij = weight connect‑
ard normal distribution curve, using the Mersenne Twister ing node i of input layer and node j of the hidden layer,
method, was used as the initial value for all the weights and wjo = weight connecting node j of hidden layer and node of
biases in this study (Matsumoto and Nishimura 1998). output layer, wj = bias at node j of the hidden layer, wo = bias
The ITS was calculated using Eqs. (3) to (5), applying the at node of output layer, f = activation function, ITS
̂ = output
initial weights and biases connecting the input layer to the value.
output layer via the hidden layer in the feedforward process The input values ( xi ) were multiplied by the weights ( wij )
shown in Fig. 6. connecting the input layer and hidden layer and then added
∑
I to the biases ( wj ) of the hidden layer to obtain a value ( hj ).
hj = wij xi + wj (3) Subsequently, the value ( hj ) was substituted into the activa‑
i=1 tion function to be converted to another value ( netj).
The sigmoid function was used on the hidden node to
( )
netj = f hj (4) make the ANN model nonlinear, as shown by Eq. (6) [27].
The sigmoid function has frequently been used to predict the
13
Arabian Journal for Science and Engineering
pavement material properties, especially for ANN structures can be halted when the absolute values of the partial deriva‑
with one hidden layer [28]. In addition, Gandhi et al. [29] tives of all the weights and biases are less than 0.01, thus
used the sigmoid function to predict the ITS of asphalt pave‑ satisfying the criterion [30]. However, the partial derivatives
ment. Therefore, the sigmoid function was selected as the of all the weights and biases for the ITS predicted through
activation function in this study. The value ( netj ) obtained the first feedforward process did not satisfy the criterion in
from the activation function was multiplied by the weights this study. Therefore, training of the ANN model was started
( wjo ) connecting the hidden layer and output layer and then using the resilient backpropagation method by, repetitively,
added to the bias ( wo ) of the output layer to finally obtain correcting each weight and bias individually until the par‑
the output value ( ITS
̂). tial derivatives of all the weights and biases satisfied the
criterion [31].
1
f (x) = (6) The learning conditions applied in the resilient back‑
1 + e−x
propagation method used in this study are listed in Table 3
The error for the ITS predicted through the feedforward [32, 33]. The learning step for the model started by adding
process against the measured ITS was calculated as the sum or subtracting the initial change in the learning amount ( 𝛥0 ),
of squared errors, as shown by Eq. (7). as listed in Table 3, to or from the initial weight and bias
that were not satisfied in the first feedforward process. Each
N ( )2
∑ initial weight and bias was increased or decreased individu‑
E= ̂n − ITSn
ITS (7)
n=1
ally, to be used as the weight and bias in the learning step,
by adding or subtracting the initial change in the learning
where E = sum of squared errors, ITS
̂ = predicted ITS (MPa), amount according to the negative or positive sign of the par‑
ITS = measured ITS (MPa). tial derivative for each initial weight and bias.
The partial derivative (∂E/∂w) for each weight and bias, The ITS was predicted again using the weights and biases
which reveals the change in error according to the change in corrected by the initial change in the learning amount in
each weight or bias, was calculated. The prediction of ITS order to recheck whether the partial derivative of each
13
Arabian Journal for Science and Engineering
13
Arabian Journal for Science and Engineering
4 Validation of ANN Model MAE is the error in the magnitude of the predicted ITS
compared to the measured ITS, while MAPE is the error
The ITSs predicted from the data of 62 training sets and in the rate of the predicted ITS compared to the measured
26 test sets, measured at 88 target sections, were plotted as ITS. In the case of the ITS predicted from the data of the
shown in Fig. 10. In addition to the coefficient of determi‑ training set, the MAE, MAPE, and coefficient of determi‑
nation, the mean absolute error (MAE) and mean absolute nation with respect to the measured ITS were calculated as
percentage error (MAPE) were calculated using Eqs. (12) 0.14 MPa, 14%, and 0.82, respectively. In the case of the ITS
and (13), respectively: predicted from the data of the test set, the MAE, MAPE, and
∑N � coefficient of determination were 0.14 MPa, 12%, and 0.79,
̂n ��
n=1 �� respectively, showing results similar to those for the train‑
ITSn − ITS
� (11)
MAE =
N ing set data. The MAEs and MAPEs were relatively low,
while the coefficients of determination were relatively high
N | for both the training and test sets. The MAPE of the training
∑ ̂n || 100
| ITSn − ITS
MAPE = | |× (12) dataset was 2% larger than that of the test dataset. However,
| ITSn | N the MAPE of the training dataset and test dataset could be
n=1 | |
considered the same because of the p value of 0.457 higher
where MAE = mean absolute error (MPa), MAPE = mean than 0.05 threshold, as a result of the t test for the absolute
absolute percentage error (%). percentage error of the training and test datasets.
13
Arabian Journal for Science and Engineering
13
Arabian Journal for Science and Engineering
13
Arabian Journal for Science and Engineering
References 19. Altun, H.; Bilgil, A.; Fidan, B.C.: Treatment of multi-dimensional
data to enhance neural network estimators in regression problems.
Expert Syst. Appl. 32, 599–605 (2007)
1. MOLIT: Road Practice Manual. Ministry of Land, Infrastructure
20. Guh, R.S.: Effects of non-normality on artificial neural network
and Transport (2019)
based control chart pattern recognizer. J. Chin. Inst. Ind. Eng.
2. MOLIT: Yearbook of Road Statistics. Ministry of Land, Infra‑
19(6), 13–22 (2002)
structure and Transport, Korea (2019)
21. Kumar, U.A.: Comparison of neural networks and regression
3. Lee, J.G.; Jeon, G.S.; Nam, M.S.; Kim, K.S.; Lee, J.S.: A Study
analysis: a new insight. Expert Syst. Appl. 29(2), 424–430 (2005)
on an Establishment of Countermeasures for Adaption of Express‑
22. Melesse, A.M.; Ahmad, S.; McClain, M.E.; Wang, X.; Lim, Y.H.:
ways to Climate. Report, Report, Expressway and Transportation
Suspended sediment load prediction of river systems: an artificial
Research Institute, Korea Expressway Corporation (2017)
neural network approach. Agric. Water Manag. 98(5), 855–866
4. KEC: Highway Pavement Maintenance System Operation Manual.
(2011)
Korea Expressway Corporation (1996)
23. Wilson, E.B.; Hilferty, M.M.: The distribution of Chi square. Proc.
5. MOLIT: Annual Research Report on National Highway Pavement
Natl. Acad. Sci. U.S.A. 17(12), 684–688 (1931)
Management System in 2018. Ministry of Land, Infrastructure and
24. Matsumoto, M.; Nishimura, T.: Mersenne twister: a 623-dimen‑
Transport, Korea (2019)
sionally equidistributed uniform pseudo-random number genera‑
6. Kennedy, T.W.; Hudson, W.R.: Application of the indirect tensile
tor. ACM Trans. Model. Comput. Simul. 8(1), 3–30 (1998)
test to stabilized materials. Highway Res. Rec. 235, 36–48 (1968)
25. Gong, H.; Sun, Y.; Mei, Z.; Huang, B.: Improving accuracy of rut‑
7. Islam, M.R.; Hossain, M.I.; Tarefder, R.A.: A study of asphalt
ting prediction for mechanistic-empirical pavement design guide
aging using indirect tensile strength test. Constr. Build. Mater.
with deep neural networks. Constr. Build. Mater. 190, 710–718
95, 218–223 (2015)
(2018)
8. Baus, R.L.; Stires, N.R.: Mechanistic-Empirical Pavement Design
26. Rousseeuw, P.J.; Hubert, M.: Anomaly detection by robust statis‑
Guide Implementation. Technical Report, No. FHWA-SC-10-01,
tics. WIREs Data Min. Knowl. Discov. 8(2), e1236 (2018)
University of South Carolina, US (2010)
27. Nair, V.; Hinton, G.E.: Rectified linear units improve restricted
9. Kim, D.H.; Lee, J.M.; Moon, K.H.; Park, J.S.; Suh, Y.C.; Jeong,
Boltzmann machines. In: Proceedings of the 27th International
J.H.: Development of remodeling index model to predict priority
Conference on Machine Learning, Haifa, Israel (2010)
of large-scale repair works of deteriorated expressway concrete
28. Abambres, M.; Ferreira, A.: Application of artificial neural net‑
pavements in Korea. KSCE J. Civ. Eng. 23(5), 2096–2107 (2019)
works in pavement management. In: Proceedings of the Interna‑
10. Choi, J.H.; Adams, T.M.; Bahia, H.U.: Pavement roughness mode‑
tional Conference on Traffic Development, Logistics and Sustain‑
ling using back-propagation neural networks. Comput. Aided Civ.
able Transport, 1-11, Opatija, Croatia, 2017 (2017)
Infrastruct. Eng. 19(4), 295–303 (2004)
29. Gandhi, T.; Xiao, F.; Amirkhanian, S.N.: Estimating indirect ten‑
11. Achanta, A.S.; Kowalski, J.G.; Rhodes, C.T.: Artificial neural
sile strength of mixtures containing anti-stripping agents using an
networks: implications for pharmaceutical sciences. Drug Dev.
artificial neural network approach. Int. J. Pavement Res. Technol.
Ind. Pharm. 21(1), 119–155 (1995)
2(1), 1–12 (2009)
12. Kim, S.; Gopalakrishnan, K.; Ceylan, H.: Neural networks appli‑
30. Günther, F.; Fritsch, S.: Neuralnet: training of neural networks. R.
cation in pavement infrastructure materials. Intell. Soft Comput.
J. 2(1), 30–38 (2010)
Infrastruct. Syst. Eng. 259, 47–66 (2009)
31. Riedmiller, M.: Advanced supervised learning in multi-layer per‑
13. KS F 2382: Standard Test Method for Indirect Tension of Asphalt
ceptions—from backpropagation to adaptive learning algorithms.
Mixtures. Korean Industrial Standards, Korean Standards Asso‑
Comput. Stand. Interfaces 16(3), 265–278 (1994)
ciation, Seoul, Korea (2013)
32. Igel, C.; Hüsken, M.: Empirical evaluation of the improved rprop
14. Gillespie, T.D.; Paterson, W.; Sayers, M.W.: Guidelines for Con‑
learning algorithms. Neurocomputing 50, 105–123 (2003)
ducting and Calibrating Road Roughness Measurements. World
33. Anastasiadis, A.D.; Magoulas, G.D.; Vrahatis, M.N.: New glob‑
Bank Technical Paper, No. WTP 46, World Bank Group, Wash‑
ally convergent training scheme based on the resilient propagation
ington, DC, US (1986)
algorithm. Neurocomputing 64, 253–270 (2005)
15. Mamlouk, M.; Vinayakamurthy, M.; Underwood, B.S.; Kaloush,
34. Giustolisi, O.; Laucelli, D.: Improving generalization of artifi‑
K.E.: Effects of the international roughness index and rut depth
cial neural networks in rainfall–runoff modelling. Hydrol. Sci. J.
on crash rates. Transp. Res. Rec. 2672(40), 418–429 (2018)
50(3), 439–457 (2005)
16. Mubaraki, M.: Highway subsurface assessment using pavement
35. Garson, G.D.: A comparison of neural network and expert systems
surface distress and roughness data. Int. J. Pavement Res. Technol.
algorithms with common multivariate procedures for analysis of
9(5), 393–402 (2016)
social science data. Soc. Sci. Comput. Rev. 9(3), 399–434 (1991)
17. MOLIT: Road Pavement Structure Design Manual. Ministry of
36. Goh, A.T.: Back-propagation neural networks for modeling com‑
Land, Infrastructure and Transport, Korea (2015)
plex systems. Artif. Intell. Eng. 9(3), 143–151 (1995)
18. Kline, R.B.: Methodology in the Social Sciences: Principles and
Practice of Structural Equation Modeling, 2nd edn. Guilford
Press, New York (2005)
13