Professional Documents
Culture Documents
net/publication/309590088
CITATIONS READS
0 933
4 authors:
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Ahmad Haj Mosa on 17 December 2018.
Abstract—This paper, presents a novel online adaptive neuro- a clearly superior performance in classification and regression
computing framework for a robust time-series forecast. The problems in various practical domains has made them to be a
proposed framework does mimic the human minds biological preferred method for time-series forecasting [2].
two-thinking model. Our mind makes decisions/calculations using
a two-connected systems. A first system, System 1, the so-called Forecasting is without any doubt a major challenge that
the intuitive system, makes decisions based on our experience. characterizes most global data analysis. That is, an accurate
A second system, System 2, the controller system, does control and timely flow of information is needed for individuals or
the decisions of System 1 by either modifying or trusting them. experts to make the relevant decisions regarding a certain
Similarity to the human minds two-systems model, this paper phenomenon under observation. It can be easier to make a
proposes an artificial framework consisting of two cellular neural
network (CNN) systems. The first CNN processor does represent decision based data or knowledge on a future evolution of
the intuitive system and we call it Intuitive-CNN. The Second the system. However, bid data can be very challenging due
CNN processor does represent the controller system, which is to the uncertainty of the related variables [3]. Thus, time-
called Controller-CNN. Both are connected within a general series forecasting will be helpful for making a calculated
framework that we name OSA-CNN. The proposed framework decision and predicting the future trend of the variables. The
is extensively tested, validated and benchmarked with the best
state-of-the-art related methods, while involving real field time- phenomena that are used to produce time-series can sometimes
series data. Multiple scenarios are considered: traffic flow data be unknown, and the information available for forecasting
extracted from the PeMs traffic database and the 111 time is by large limited to the past values of the time-series [4].
series collected from the so-called NN3 competitions. The novel Therefore, it is essential and vital to utilize the most relevant
OSA-CNN concept does remarkably higly outperform the state- number of previous values, termed lag, when undertaking a
of-the-art competing methods regarding both performance and
universality. time-series forecasting. This is the gap that ANN is mostly
developed to close.
Index Terms—Time series forecast (TFS), Cellular neural This paper is organized as follows, Section II address the
networks (CNN), Echo state network (ESN), Reference neural
network adaptive control (MRNNAC), Recursive particle swarm related work . Section III provides background information
optimization (RPSO), Online self adaptive CNN (OSA-CNN) related to the techniques that are involved in our method.
Section IV presents OSA-CNN models. Section V explains
the learning algorithm of OSA-CNN. Section VI shows the
I. I NTRODUCTION
obtained results using different two benchmark online data
The importance of time-series in science and technology sets. Finally a conclusion is given in Section VII.
is extremly high nowadays. It is a fact that that time-series
forecasting (TSF) is widely utilized in several areas such as II. R ELATED W ORKS
engineering, science, and finances [1]. The high interest for Compared to earlier works especially to some of the existing
time-series analysis and forecasting is motivated by the need statistical forecasting methods, neural network approaches
for understanding and predicting the future of various tech- have proven several outstanding characteristics that include
nical, social, and natural systems. In this perspective several the following ones: (a) artificial neural networks can use both
forecasting methods have been developed. Some of which rely linear and nonlinear data; (b) artificial neural networks are
on either linear or nonlinear statistical models. One of these non-parametric, that is, ANN does not need an explicit un-
promising forecasting concepts do involve artificial neural net- derlying model if compared to other forecasting methods; and
works (ANN). Over the past decades, time-series forecasting (c) ANN is very flexible and universal and hence can be used
has been predominately implemented using statistical analysis for more complicated and complex data structures/models.
based methods. One prominent example is the autoregressive The above listed characteristics make ANN to be regarded as
integrated moving averages (ARIMA) which has been used the best current methods/concepts for time-series forecasting.
by scientists for several years. However, regarding both the Several empirical studies have shown a significantly superior
emerging technologies and the high complexity of much recent performance of ANN over other statistical forecasting tools.
data, artificial neural networks have been preferred and have The superiority measure/benchmarking has been mainly based
become a serious processing alternative in time-series forecast- on a single or a small set of time-series. A special and very
ing. The fact that artificial neural networks are characterized by prominent feature of ANN is its ability to combine both
Manuscript received December 1, 2012; revised August 26, 2015. Corre- long and short-term forecasting approaches to provide a better
sponding author: M. Shell (email: http://www.michaelshell.org/contact.html). system-model with highest performance potential [5].
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2
Real Historical a) We take each two samples of the original series, say
Values vk0 (1) and vk0 (2), then we interpolate them with a square
pulse that has the length Λ (interpolation size or the time-
distance between two samples of the original time-series)
Controller Predicted which starts at vk0 (1) and ends at vk0 (2).
System Values
Intuitive b) This square pulse has the constant value of vk0 (1) over all
Control System the interval of size Λ between lag-time (1) and lag-time
Action (2).
Recent
Predictions The aim of this interpolation process is to make the system
Time (driven by the controller part) doing several prediction trials of
Delay the same target and eventually converges at the best possible
prediction value. The resulted signals (i.e square pulses) are
Fig. 3. The OSA-CNN Framework. The Intuitive-System takes the historical denoted by vsk0 (ti) and vsk1 (ti) to vskh (ti) respectively.
values to predict the future value of a time series. The Controller-System reads Note that each of this signals has the length of Λ. In the
and compares the historical values with the corresponding prediction values next step, as in MRNNAC, we do need to obtain the reference
and manipulate the Intuitive-System output to improve the future performance
series of the pulses series. A reference series is the same as the
pulses series; however the difference is that the reference series
variant set-point value is valued with its correspond- consist of smooth steps instead of sharp square steps. The
ing reference model to train the controller. In this resulted smooth path (see vrj (ti) in Equation (3)) mimics the
training phase, only the controller network weights desired path that the plant should follow. The transformation
need to be learned while the process network weights from squares into smooth steps is generally done by using the
are fixed from the first training phase (training of the following reference model, which is very known in the related
proceed network). literature[20], [22], [23], [24], [25]:
d2 vrj (ti) vrj (ti)
2
= −αvrj (ti) + −ρ + γvsj (ti) (3)
IV. OSA-CNN dti dti
OSA-CNN is built by integrating: a) a continuous-time where j = k0 to kh , representing the original and lagged
nonlinear CNN oscillator system; b) the echo-state property square steps series; α, ρ and γ are the feedback, damping and
and the easy training approach of ESN; and c) the theory of the control coefficients that characterize the shape of the desired
two-systems’ model of thinking together into one framework. smooth path.These coefficients are selected empirically. The
This combination of advantages makes OSA-CNN to be a selected values are presented in the evaluation Section VI.
precious architecture. As it is presented in Fig. 3, OSA-CNN Equation (3) is a second order differential equation that
has two connected systems, mimicing the Two-Systems model needs to be solved. Given a sequence/series of square pulses
of human cognitive processes. System 1 is the controller vsj (ti) and zero initial condition (at the very beginning oft he
model (Controller-System). System 2 is the process model sequence) we solve (3) using the Matlab [26] Adams-Moulton
(Intuitive-System). Both systems are integrated similarly to solver [27] (ode113) solver with a variable step size between
MRNNAC. The role of the Intuitive-System is to predict the 0.05 and 0.1. Accordingly, the solution gives the reference
future value of a time-series given its historical records. The smooth step series vrj (ti). Eventually, we get a set of two
role of the Controller-System is then to assess the recent time-series which is used in the training and testing processes
prediction performance of the Intuitive-System by comparing of OSA-CNN. These time series are:
it with the corresponding real values. Accordingly it controls a) The OSA-CNN input (square pulse) signals
the Intuitive-System to improve the future performance. In vsk1 (ti), . . . vskh (ti): the step series of the historical
other words, the Controller-System drives the Intuitive-System values that are used as inputs for the prediction model.
in such a way that the recent prediction error is minimized. b) The OSA-CNN output (smooth pulse) signals vrk0 (ti):
In the following, we address the data preparation needed for the reference series of the future values that need to be
OSA-CNN along with a review of modeling details. predicted.
The output of OSA-CNN is denoted by vr ˆ k0 (ti). The task of
A. OSA-CNN Data Preparation OSA-CNN is to minimize the difference between the input
signals vsk1 (ti). . . . vskh (ti) and the history of the recent
Given a time series v(t), the forecasting goal is to predict prediction values vr ˆ k1 (ti), . . . vr
ˆ kh (ti).
v(t) while considering the h historical values v(t − k1 )...v(t −
kh ) as shown in Fig 4(a), where ki is a delay or lag. For
modeling purposes, we treat the original and lagged series B. OSA-CNN Model
as separate quantities denoted by vk0 (t) and vk1 (t) to vkh (t) As it is mentioned above, OSA-CNN consists of two
respectively. For modeling purposes, the original and lagged connected systems the Controller-System and the Intuitive-
time-series are transformed into a series of quadratic pulses System. Each one of this two systems consists of a hidden and
(i.e. step-functions) wich are delayed by one time-lag of the output layer. The hidden layer is an ordinary cellular neural
time series and which look as following (see Fig 4(b)): network as presented in Sub-section III-B. The output layer is
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 5
Linear
+ +
Linear
Linear +
+ +
PSO
C-CNN
C-LR
I-CNN
-
I-LR +
+
Satlins Satlins
+ ∫ + ∫
-1 -1
+ +
Fig. 5. The OSA-CNN four layers model. Two modes are presented. The online operating (solid line connections) and offline training (doted line) modes.
In the operating mode. The first layer is the Controller CNN (C-CNN) connected to the Controller linear layer C-LR. The third layer is the Intuitive CNN
(I-CNN) connected to the Intuitive linear layer I-LR. The controller output u(ti) adapts the I-LR biases with respect to the difference between the real
historical values vsk1 (ti)..vskh (ti) and the corresponding predicted values vr
ˆ k1 (ti)..vr
ˆ kh (ti). In addition, four blocks visualizing the models of C-CNN
cell state, I-CNN cell state, C-LR and I-LR. In the offline training mode, PSO reads the prediction error and adapt accordingly the the C-LR weights
nothing but a linear layer. Accordingly, OSA-CNN consists of of the I-CNN outputs. In addition, the control action for C-
four connected layers: Controller CNN (C-CNN), Controller LR to I-LR is realized as an adaptive bias added to weighted
Linear Regression (C-LR), Intuitive CNN (I-CNN) and In- sum in I-LR. In this way, the Controller-System adapts the
tuitive Linear Regression (I-LR) models. Figure 5 presents Intuitive-System output by regulating the biases of the I-LR
these four layers. As it is shown, the C-CNN is fed by the h layer. In the following, we do address the modeling equation
lag reference square-pulses vsk1 (ti)..vskh (ti) representing the of both the first Controller-System and the Intuitive-CNN.
recent real h values. It is also fed by the h lag predicted values Equations (4) and (5) (6) give the state, output models of
vr
ˆ k1 (ti)..vr
ˆ kh (ti). The last input for C-CNN is the Controller-
System output feedback u(ti). C-CNN feeds forward the linear
layer C-LR which does a liner mapping (weighted sum) to
the Controller-System output. The I-CNN has two inputs, first
one is again the h lag reference square pulse vsk1 (ti)..vskh (ti)
and the Intuitive-System output feedback vr ˆ k0 (ti). The I-CNN
feeds forward the linear layer I-LR which does a weighted sum
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 6
W c-cnn
2 = (ω c-cnn c-cnn
21,1 . . . ω 2n,h ), W 3
c-cnn
= (ω c-cnn c-cnn
31,1 . . . ω 3n,h )
c-cnn c-cnn c-cnn
and W 4 = (ω 41,1 . . . ω 4n,n ) are the C-CNN control
templates; W c-lri = (ω c-lr c-lr
i,1 . . . ω n,i ) is the ith C-LR template;
c-cnn
ε = (ε1 . . . εn ) and ε = (εc-lr
c-cnn c-cnn c-lr c-lr
1 . . . εn ) are biases.
While, Equations (7), (8) and (9) give the state, output
models of the I-CNN and I-LR respectively:
n
dxi-cnn (ti) X
i
= −xi-cnn
i (ti) + ω i-cnn i-cnn
1i,j yj (ti) + (7)
dti j=1
h
X
ω i-cnn i-cnn
2i,j vskj (ti) + ω 3i vr ˆ k0 (ti) + εi-cnn
i
(a)
j=1
1 i-cnn
yii-cnn (ti) = (|x + 1| − |xi-cnn − 1|) (8)
2 i i
n
(ω i-lr i-cnn
P
vr
ˆ k0 (ti) = i yi (ti) + ui (ti)) (9)
i=1
V. OSA-CNN L EARNING
C-CNN and C-LR respectively: The training of OSA-CNN is subjected to three main
n phases. First, the C-CNN and I-CNN feedback and control
dxc-cnn (ti) X
i
= −xc-cnn
i (ti) + ω c-cnn c-cnn
1i,j yj (ti) + (4) templates and the biases are generated randomly. The random
dti j=1 template generators must ensure the echo state and CNN
h
X h
X stability properties. In the second step, the I-LR is trained
ω c-cnn
2i,j vskj (ti) + ω c-cnn
3i,j vr
ˆ kj (ti) + using a regularized least square method. In the last phase,
j=1 j=1 C-LR templates are trained using recursive particle swarm
n
X optimization (RPSO). In the following, we explain the three
ω c-cnn c-cnn
4i,j uj (ti) + εi training phases.
j=1
TABLE I
T HE RPSO CONFIGURATION PARAMETERS . T HESE CONFIGURATIONS ARE
USED IN ALL OF THE STUDIED TSF CASES
Parameter Value
Swarm Size 24
M AXIT ER 100
$1 0.9
$2 0.4
δ1,1 2.5
δ1,2 0.5
δ2,1 2.5
δ2,2 0.5 (a)
Stopping Criteria 1 MAXITER is exceed
Stopping Criteria 2 Gmin = 1e − 10
Stopping Criteria 3 Θ = 10
(a) (b)
(c) (d)
(e) (f)
Fig. 8. The OSA-CNN Traffic flow prediction using the detector 1006210 from PEMS data set. It presents OSA-CNN predicted values (in blue) compared
with the ground truth traffic flow (in red) of the first two days of November 2009 (1st is Sunday (Weekend) and the 2nd is a Monday (Working day)) using
different aggregation levels
prediction techniques combined with his proposed intra- Fig.7(a),7(b) where OSA-CNN outputs drive I-CNN to a
day trend model. For comparison purposes, we choose better performance. In addition, Figures 8(a), 8(b), 8(c)
the best five methods. These methods are: Auto Regres- and 8(d) show the OSA-CNN predicted values compared
sive Moving Average Model (ARMA) [46] [47], Auto with the ground truth traffic flow of the first two days of
Regressive Integrated Moving Average Model (ARMA) November 2009 (1st is Sunday (Weekend) and the 2nd is
[48][49], Support Vector Regression [50] [51], Bayesian a Monday (Working day)). OSA-CNN shows a capability
Network [52] and Radial Basis Neural Network [53]. of reliable prediction in both weekends and working days.
The evaluation is done over the aggregation levels 3- b) In the second evaluation context, both I-CNN and OSA-
min, 5-min, 10-min and 15-min. The related comparison CNN are tested for long-term (30-min and 60-min) one-
with I-CNN and OSA-CNN is presented in Tables ??, step prediction. As in the short-term prediction case,
IV, V, VI. The general conclusion is I-CNN does over- OSA-CNN did improve I-CNN outputs and gave a perfor-
perform the related methods by 0.14 to 4.08 % MRE. mance consistent with the short-term forecast one. Table
What is more interesting is that OSA-CNN does improve VII presents the related performance values and Figures
I-CNN performance by 0.8 to 4 % MRE. The comparison 8(e), 8(f) shows the prediction graph of the same time
between OSA-CNN and I-CNN is more visible in Fig.9. period like for the previous context.
An example of OSA-CNN control actions can be seen in c) In the last context, we test OSA-CNN to do a multi-
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 11
TABLE IX
T HE NUMBER NN3 FORECASTING COMPETITION TIME SERIES WITH
RESPECT TO THE LENGTH AND SEASONALITY OF EACH TIME SERIES
TABLE III
T HE PERFORMANCE OF OSA-CNN AND I-CNN VERSUS OTHER MODELS , WITH 3- MIN AGGREGATION LEVEL . DATA USED IN THIS EVALUATION IS THE
MEASURED TRAFFIC DATA PROVIDED BY THE P E MS DATASET OF THE DETECTOR 1006210. DATA ARE COLLECTED BETWEEN O CTOBER 1, 2009 AND
N OVEMBER 30, 2009. T HE RESULTS OF OTHER METHOD IS TAKEN FROM [41]
TABLE IV
T HE PERFORMANCE OF OSA-CNN AND I-CNN VERSUS OTHER MODELS , WITH 5- MIN AGGREGATION LEVEL . DATA USED IN THIS EVALUATION IS THE
MEASURED TRAFFIC DATA PROVIDED BY THE P E MS DATASET OF THE DETECTOR 1006210. DATA ARE COLLECTED BETWEEN O CTOBER 1, 2009 AND
N OVEMBER 30, 2009. T HE RESULTS OF OTHER METHOD IS TAKEN FROM [41]
TABLE V
T HE PERFORMANCE OF OSA-CNN AND I-CNN VERSUS OTHER MODELS , WITH 10- MIN AGGREGATION LEVEL . DATA USED IN THIS EVALUATION IS THE
MEASURED TRAFFIC DATA PROVIDED BY THE P E MS DATASET OF THE DETECTOR 1006210. DATA ARE COLLECTED BETWEEN O CTOBER 1, 2009 AND
N OVEMBER 30, 2009. T HE RESULTS OF OTHER METHOD IS TAKEN FROM [41]
TABLE VI
T HE PERFORMANCE OF OSA-CNN AND I-CNN VERSUS OTHER MODELS , WITH 15- MIN AGGREGATION LEVEL . DATA USED IN THIS EVALUATION IS THE
MEASURED TRAFFIC DATA PROVIDED BY THE P E MS DATASET OF THE DETECTOR 1006210. DATA ARE COLLECTED BETWEEN O CTOBER 1, 2009 AND
N OVEMBER 30, 2009. T HE RESULTS OF OTHER METHOD IS TAKEN FROM [41]
TABLE VII
T HE PERFORMANCE OF OSA-CNN AND I-CNN WITH 30- MIN AND 60- MIN AGGREGATION LEVEL . DATA USED IN THIS EVALUATION IS THE MEASURED
TRAFFIC DATA PROVIDED BY THE P E MS DATASET OF THE DETECTOR 1006210. DATA ARE COLLECTED BETWEEN O CTOBER 1, 2009 AND N OVEMBER
30, 2009
)
N
no an
16,55
au
)
es
(G
[17] H. Jaeger, “The ?echo state? approach to analysing and training recur-
ko
re
el
ya
16,13
T
m
)
io
m
ss
he
e
r i
o
e
)
lg
R
N
A
15,18
i
R
et
em
(G
en
an
ES
(E
14,95
s
s
he
llie
N
o
Fl
N
C
14,84
I
)
adaptive controller,” ISA transactions, vol. 37, no. 1, pp. 21–39, 1998.
M
SS
14,56
(A
)
N
l
EN
i
W
N
m
-C
ah
SMAPE Theory and Applications, vol. 18, no. 2, pp. 187–197, 1976.
[22] F. Rossomando, C. Soria, D. Patiño, and R. Carelli, “Model reference
Fig. 13. The comparison between OSA-CNN against the related state of the adaptive control for mobile robots in trajectory tracking using radial
art methods over the 111 time series of NN3 time series competition. OSA- basis function neural networks,” Latin American applied research,
CNN proved to be an robust TSF algorithm with 13.18 % comparing to other vol. 41, no. 2, pp. 177–182, 2011.
methods [23] S. Wei, “A model reference-based adaptive pid controller for robot
motion control of not explicitly known systems,” International Journal
of Intelligent Control and Systems, vol. 12, no. 3, pp. 237–244, 2007.
VII. C ONCLUSION [24] S. G. Shuzhi, T. H. Lee, and C. J. Harris, Adaptive neural network
control of robotic manipulators. World Scientific, 1998, vol. 19.
[25] H. A. Kingravi, G. Chowdhary, P. A. Vela, and E. N. Johnson, “Repro-
The conclusion goes here. ducing kernel hilbert space approach for the online update of radial bases
in neuro-adaptive control,” Neural Networks and Learning Systems,
IEEE Transactions on, vol. 23, no. 7, pp. 1130–1141, 2012.
[26] MATLAB, version 7.14.0 (R2013b). Natick, Massachusetts: The
R EFERENCES MathWorks Inc., 2013.
[27] E. Hairer, S. NORSETT, and G. Wanner, Solving Ordinary, Differential
[1] W. Yan, “Toward automatic time-series forecasting using neural net- Equations I, Nonstiff problems/E. Hairer, SP Norsett, G. Wanner, with
works,” Neural Networks and Learning Systems, IEEE Transactions on, 135 Figures, Vol.: 1. 2Ed. Springer-Verlag, 2000, 2000.
vol. 23, no. 7, pp. 1028–1039, 2012. [28] B. Mirzai, Z. Cheng, and G. S. Moschytz, “Learning algorithms for
[2] Y. Lv, Y. Duan, W. Kang, Z. Li, and F.-Y. Wang, “Traffic flow prediction cellular neural networks,” in IEEE International Symposium On Circuits
with big data: A deep learning approach,” Intelligent Transportation And Systems. INSTITUTE OF ELECTRICAL ENGINEERS INC
Systems, IEEE Transactions on, vol. 16, no. 2, pp. 865–873, 2015. (IEEE), 1998, pp. 159–162.
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 15
[29] A. Fasih, J. C. Chedjou, and K. Kyamakya, “Cellular neural network [54] S. Makridakis, C. Chatfield, M. Hibon, M. Lawrence, T. Mills, K. Ord,
trainer and template optimization for advanced robot locomotion, based and L. F. Simmons, “The m2-competition: A real-time judgmentally
on genetic algorithm,” in Mechatronics and Machine Vision in Practice, based forecasting study,” International Journal of Forecasting, vol. 9,
2008. M2VIP 2008. 15th International Conference on. IEEE, 2008, no. 1, pp. 5–22, 1993.
pp. 317–322. [55] S. F. Crone, M. Hibon, and K. Nikolopoulos, “Advances in forecasting
[30] S. SanthoshKumar, J. Vignesh, L. Rangarajan, V. Narayanan, K. Ran- with neural networks? empirical evidence from the nn3 competition on
garajan, and A. Venkatkrishna, “A fast time scale genetic algorithm time series prediction,” International Journal of Forecasting, vol. 27,
based image segmentation using cellular neural networks (cnn),” in no. 3, pp. 635–660, 2011.
Signal Processing and Communications, 2007. ICSPC 2007. IEEE [56] N. K. Ahmed, A. F. Atiya, N. El Gayar, and H. El-Shishiny, “A com-
International Conference on. IEEE, 2007, pp. 908–911. bined neural network/gaussian process regression time series forecasting
[31] H.-L. Wei and S. A. Billings, “Generalized cellular neural networks (gc- system for the nn3 competition.”
nns) constructed using particle swarm optimization for spatio-temporal [57] G. D?yakonov Alexander, “Simple knn-method for times series predic-
evolutionary pattern identification,” International journal of Bifurcation tion.”
and Chaos, vol. 18, no. 12, pp. 3611–3624, 2008. [58] H. Chen and X. Yao, “Ensemble regression trees for time series
[32] T.-J. Su, J.-C. Cheng, and Y.-D. Sun, “Particle swarm optimization predicitions,” methods, vol. 11, p. 6, 2007.
with time-varying acceleration coefficients based on cellular neural [59] P. Flores, C. Anaya, H. M. Ramı́rez, and L. B. Morales, “Automated
network for color image noise cancellation,” in ICDT 2011, The Sixth linear modeling of time series with self adaptive genetic algorithms,”
International Conference on Digital Telecommunications, 2011, pp. in Neural Networks, 2007. IJCNN 2007. International Joint Conference
109–115. on. IEEE, 2007, pp. 1389–1396.
[33] P. A. Mastorocostas, “Resilient back propagation learning algorithm for [60] I. Ilies, H. Jaeger, O. Kosuchinas, M. Rincon, V. Sakenas, and N. Vaske-
recurrent fuzzy neural networks,” Electronics Letters, vol. 40, no. 1, pp. vicius, “Stepping forward through echoes of the past: forecasting with
57–58, Jan 2004. echo state networks,” Short report on the winning entry to the NN3
[34] S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural financial forecasting competition, available online at http://www. neural-
Computation, vol. 9, no. 8, pp. 1735–1780, Nov 1997. forecasting-competition. com/downloads/NN3/methods/27-NN3 Herbert
Jaeger report. pdf, 2007.
[35] Y. Shi and R. Eberhart, “A modified particle swarm optimizer,” in
[61] S. B. Taieb, G. Bontempi, A. F. Atiya, and A. Sorjamaa, “A review and
Evolutionary Computation Proceedings, 1998. IEEE World Congress on
comparison of strategies for multi-step ahead time series forecasting
Computational Intelligence., The 1998 IEEE International Conference
based on the nn5 forecasting competition,” Expert systems with appli-
on. IEEE, 1998, pp. 69–73.
cations, vol. 39, no. 8, pp. 7067–7083, 2012.
[36] Y. Shi and R. C. Eberhart, “Empirical study of particle swarm optimiza-
tion,” in Evolutionary Computation, 1999. CEC 99. Proceedings of the
1999 Congress on, vol. 3. IEEE, 1999.
[37] A. Ratnaweera, S. Halgamuge, and H. C. Watson, “Self-organizing
hierarchical particle swarm optimizer with time-varying acceleration
coefficients,” Evolutionary Computation, IEEE Transactions on, vol. 8, Ahmad Haj Moas received his B.S. degree in
no. 3, pp. 240–255, 2004. mechatronics engineering from Tishreen University,
[38] G. Fornarelli and A. Giaquinto, “Adaptive particle swarm optimization Lattakia, Syria, in 2008. He received his M.S. degree
for {CNN} associative memories design,” Neurocomputing, vol. 72, in health care IT from from Carinthia University of
no. 16?18, pp. 3851 – 3862, 2009, financial Engineering Computational Applied Science, Klagenfurt, Austria, in 2012 and
and Ambient Intelligence (IWANN 2007). [Online]. Available: is currently a Ph.D. candidate, a university assistant,
http://www.sciencedirect.com/science/article/pii/S0925231209001684 and a lecturer at the Institute for Smart System Tech-
[39] Caltrans, “Performance measurement system (pems).” nology (IST), Alpen-Adria University of Klagenfurt
[40] NN3, “Time series forecasting competition for neural network and in Austria. His research interests include machine
computational intelligence.” vision, machine learning, applied mathematics, and
[41] C. Chen, Y. Wang, L. Li, J. Hu, and Z. Zhang, “The retrieval of intra- neurocomputing. He developed a variety of methods
day trend and its influence on traffic prediction,” Transportation research in the scope of human-machine interaction and pattern recognition. He has
part C: emerging technologies, vol. 22, pp. 103–118, 2012. participated in several projects the field of transportation informatics, and the
[42] L. Li, X. Su, Y. Wang, Y. Lin, Z. Li, and Y. Li, “Robust causal outcomes were patented
dependence mining in big data network and its application to traffic flow
predictions,” Transportation Research Part C: Emerging Technologies,
vol. 58, pp. 292–307, 2015.
[43] C. Tofallis, “A better measure of relative prediction accuracy for model
selection and model estimation,” Journal of the Operational Research Kyandoghere Kyamakya obtained the Ingnieur
Society, vol. 66, no. 8, pp. 1352–1362, 2014. Civil degree in Electrical Engineering in 1990 at
[44] E. L. Lehmann and G. Casella, “Theory of point estimation (springer the University of Kinshasa, DR Congo. In 1999
texts in statistics),” 1998. he received his Doctorate in Electrical Engineering
[45] R. Geary, “The ratio of the mean deviation to the standard deviation as at the University of Hagen in Germany. He then
a test of normality,” Biometrika, vol. 27, no. 3/4, pp. 310–332, 1935. worked three years as post-doctorate researcher at
[46] W. A. Fuller, Introduction to statistical time series. John Wiley & Sons, the Leibniz University of Hannover in the field of
2009, vol. 428. Mobility Management in Wireless Networks. From
[47] P. Whitle, Hypothesis testing in time series analysis. Almqvist & 2002 to 2005 he was junior professor for Positioning
Wiksells, 1951, vol. 4. and Location Based Services at Leibniz University
[48] M. S. Ahmed and A. R. Cook, Analysis of freeway traffic time-series of Hannover. Since 2005 he is full Professor for
data by using Box-Jenkins techniques, 1979, no. 722. Transportation Informatics and Director of the Institute for Smart Systems
[49] T. Mills, “Time series techniques for economistscambridge university Technologies at the University of Klagenfurt in Austria. He is actively con-
press,” Cambridge et al, 1990. ducting research involving modeling, simulation, and test-bed evaluations for
a series of concepts in the frame of the application of information and commu-
[50] W. K. Härdle, M. Müller, S. Sperlich, and A. Werwatz, Nonparametric
nication technology in transportation. Three main transportation systems are
and semiparametric models. Springer Science & Business Media, 2012.
addressed: (a) Intelligent Transportation Systems; (b) Intelligent Vehicles and
[51] K. Takezawa, Introduction to nonparametric regression. John Wiley &
Mobile Robotics Systems; and (c) Intelligent Logistics and Supply Chains. In
Sons, 2005, vol. 606.
the research addressing these systems, a series of fundamental and theoretical
[52] A. Darwiche, Modeling and reasoning with Bayesian networks. Cam- tools from the fields of applied mathematics, electronics, and computer science
bridge University Press, 2009. are either extensively exploited or are source of inspiration for innovative
[53] W. Yang, D. Yang, Y. Zhao, and J. Gong, “Traffic flow prediction based solutions and concepts: nonlinear dynamics and synchronization, cellular
on wavelet transform and radial basis function network,” in Logistics neural networks, nonlinear image processing, cellular-neural-networks-based
Systems and Intelligent Management, 2010 International Conference on, analog computing, systems science, and computational intelligence
vol. 2. IEEE, 2010, pp. 969–972.
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 16