You are on page 1of 7

Novel Deep Learning Model with CNN and Bi-Directional

LSTM for Improved Stock Market Index Prediction


Jithin Eapen Abhishek Verma Doina Bein
Department of Computer Science Department of Computer Science Department of Computer Science
California State University New Jersey City University California State University
Fullerton, California 92831 Jersey City, NJ 07305 Fullerton, California 92831
jithin.john@csu.fullerton.edu averma@njcu.edu dbein@fullerton.edu

Abstract— Predicting variations in stock price index has been concepts. Thereby deep learning models can achieve high
an important application area of machine learning research. Due accuracy on prediction and classification tasks.
to the non-linear and complex nature of the stock market making
predictions on stock price index is a challenging and non-trivial Once an accurate model is learned from training data,
task. Deep learning approaches have become an important users will be able to feed temporal stock price data of fixed
method in modeling complex relationships in temporal data. In length to the model and be able to get the output as the
this paper: (i) we propose a novel deep learning model that prediction for the next seven days. It can be used to predict
combines multiple pipelines of convolutional neural network and short term price changes, by various stakeholders like
bi-directional long short term memory units. (ii) Proposed model investors, financial analysts, investment bankers, as one of the
improves prediction performance by 9% upon single pipeline factors affecting their investment decisions. Our primary
deep learning model and by over a factor of six upon support objective is to analyze a sequence of time-series data and give
vector machine regressor model on S&P 500 grand challenge predicted price for the next seven days. For this purpose, we
dataset. (iii) We illustrate the improvement in prediction selected the Standard and Poor’s (S&P) 500 stock dataset,
accuracy while minimizing the effects of overfitting by presenting publicly available on Yahoo Finance [2].
several variations of multiple and single pipeline deep learning
models based on different CNN kernel sizes and number of bi- The paper is organized as follows. In Section II we survey
directional LSTM units. several representative techniques of machine learning and
neural networks that are used for stock price prediction. In
Keywords— deep learning; Bi-directional LSTM; stock market Section III we describe the dataset used for benchmarking
prediction; CNN; S&P 500. learned models. In Section IV we present our research
methodology and proposed novel model, which gives
I. INTRODUCTION improved performance in comparison with the state of the art
Stock markets are some of the most import financial models. Section V presents experimental environment.
institutions of any capitalist economy. It allows enterprises to Experimental results are presented in Section VI. Concluding
raise money from the investing public to fund their growth and remarks and future work is discussed in Section VII.
operations. At the same time there is a large amount of risk
involved for the investors due to its volatile nature. Hence II. RELATED WORK
both private investors and financial institutions like banks, A. Automation of Stock Market
have always attempted to devise ways to predict the changes
occurring in the stock prices and make buy and sell decisions Stock market traders today use electronic exchange
early. systems to automate the trading of stocks based on theoretical
buy and sell prices, which are calculated based on the current
The developments in the field of machine learning allowed market price of the stock [3]. Such an automated system
the use of computers in the task predicting financial markets, provides the advantages of better response times to market
by automating the existing statistical methods of data analysis changes, more accurate trading operations, and reduced risk of
[1]. Machine learning is about building computer systems or loss due to repeated or mistaken operations.
programs that can learn from data. Traditional machine
learning approaches to stock prediction have focused on Traders have developed multiple methods for predicting
improving their performance with different techniques for future stock prices: Fundamental stock analysis is the study of
feature extraction to select the most promising features from a the business value of the stock, by considering factors like
dataset. The most popular is the support vector machine, a revenue, operating expenses, assets, products, management,
type of kernel based supervised machine learning algorithm. competitors, economic environment etc. Technical analysis is
While being versatile, they are not as powerful as neural based on the belief that markets tend to move in cycles and
networks, especially the modern deep learning networks. Deep accurate predictions can be made by studying the price
learning networks are powerful machine learning algorithms patterns and price trends of the past. Traders have been using
that make use of many cascading layers in order to learn statistical tools to analyze price and volume data. Recent
multiple levels of representations. This corresponds to advances in computing and machine learning has enabled
different levels of abstraction. The levels form a hierarchy of them to analyze larger datasets.
Figure 1. S&P 500 index closing prices from January 2, 2008 to November 27, 2018. [2]

B. Support Vector Machine Regressor output prediction fromr the previous term. This architecture
An important characteristic of stock prices is their time conforms perfectly with the time-series nature of stock price
series dependency, i.e., stock price at a particular time is data. A hierarchy of Deep RNNs have been shown to perform
dependent on the price during the previous instance. In [4] the well for prediction tasks as related to time-series data [8].
authors proposed to use Support Vector Machine (SVM) One drawback to using the RNNs is that they result in long
Regression based model to predict stock prices, as it is a connections between the final output layer, and hidden layers
suitable learning technique for recognizing patterns in a time of previous data in the sequence, thus resulting in the problem
series dataset. Different windowing methods were used to feed of the vanishing gradients. To overcome this problem,
more reliable inputs to the regression model. The accuracy of mmodifications were made to the RNN units to include gates
the model depended on choosing the best parameters for the and memory units to remember the gradients. The most
support vector machine. This pre-processing of the input data popular such units are Long Short Term Memory (LSTM)
for rectangular window and flattened window were found to units [10]. LSTM units are designed to enhance the ability of
have good prediction for stock prices 1 and 5 days ahead RNNs to remember previous values and prioritize learning
respectively. This model was applied only to the time series from newer data in a sequence upon older data. The memory
dataset of a single company in the stock market. of LSTM units also solves the problem of vanishing gradients,
C. Artificial Neural Network as the memory allows a gradient to move from one hidden
layer to the next without being reduced, letting early layers of
Artificial neural network (ANN) models have been a neural network train almost as well as later layers [9].
compared to other statistical models [5] and generally ANN LSTMs were shown to be more effective than conventional
models give better result [6]. The authors then compare RNNs [10].
various types and hybrids of ANN and multi-layer perceptron
models and found that hybrid adaptive neural networks need Bi-directional Recurrent Neural Networks [11] are used to
to be remodeled when training data changes for the time-series learn from both the forward and backward time dependencies.
forecast. In addition, each layer has to be controlled by In a Bi-directional LSTM, each unit is split into two units
reliability tests, as it is the input from the following layers of having the same input and connected to the same output. One
regression model. unit is used for the forward time sequence and the other for the
backward sequence. The common output does an arithmetic
D. Deep Neural Networks operation like sum, average or concatenation on both the units.
Recurrent Neural Networks (RNN) are a special form of Bi-directional LSTM is thus useful when learning from long
ANN where the weights computed during training are spanning time-series data.
influenced not only by the current training example, but by
previous decisions made on earlier training data. They have A technique that is suitable for feature extraction on high-
been specifically designed to deal with sequential or temporal dimensional data, as found in stock prices, is convolutional
data [7]. In RNNs, the hidden layer receives input not only neural networks [12]. A convolutional neural network (CNN)
from its immediate predecessor (input) layer, but also from a is very similar to an ANN but makes specific assumptions
previous output layer. Thus, for each input term in the about the input data. That allows it achieve higher invariance
sequential data, the weights calculated also depend on the when encoding properties of input data into the network
Output Layer

Figure 2. Diagram of basic structure of single pipeline CNN based Bidirectional LSTM network. We use this in our experiments as a baseline deep learning model
to compare with proposed novel multiple pipeline deep learning model.

architecture [13]. A typical CNN includes convolution layers, index is widely regarded as the best single gauge of large-cap
rectified linear unit (ReLU) as activation function, pooling, U.S. equities. Thereby it is an excellent measure of the
and fully connected layers. CNNs are the main workhorse to performance of overall stock market.
solve modern computer vision and deep learning problems
ever since AlexNet achieved first place in the ILSVRC 2012, S&P 500 dataset is publicly available on Yahoo Finance
an annual ImageNet Challenge [14]. CNNs need to be trained [2]. It consists of historical daily opening and closing stock
on large amounts of training data to generate deep learning prices along with a few other additional market parameters.
models that achieve high generalization accuracy. Fig. 1 shows the daily closing price of S&P 500 index from
Generalization accuracy refers to the performance of the January 2, 2008 to November 27, 2018. It has the daily data
models on test data. The hidden units, instead of merely for five days in a week for a period of 10 years. It consists of
calculating a function on the input, performs a convolution columns like daily Opening price, Closing price, High and
operation using a filter of weights. It is thus capable of Low prices, and daily Volume of stocks traded. The S&P500
exploiting the time structure of the input by performing was chosen due to its stability compared to individual
convolutions on overlapping windows. They also use pooling company stock prices. Statistical analysis shows that the
operations, which combines nearby values in input or feature autoregressive moving average model for S&P 500
space through a max, average or histogram operator. The outperforms the London Stock Exchange, suggesting better
purpose of pooling is to achieve invariance to small local potential for predictive models [16].
distortions and reduce the dimensionality of the feature space. Stock index data is segmented into sequences of 50 closing
Most deep learning architectures are created using a single prices, with the last price of each sequence being one day after
sequence of multiple layers, each performing a specific the sequence before it. We construct multiple sequences of
operation on the data. A multi-pipeline architecture, consisting length 50, and feed the deep learning model the entire batch as
of four parallel sequences of layers, has been shown to give an input. Each sequence is incremented by one day.
lower classification errors on sentiment analysis in the text Additionally, tests were performed using sequences of length
domain as compared to a single pipeline model [15]. Each 14, 28, and 56, but the length of 50 performed the best out of
pipeline was created using a sequence of CNN, Batch the three other lengths. Each of the data segments was
Normalization, ReLU activation, Pooling, and LSTM layers. normalized by subtracting each price in the sequence by the
The output of each pipeline is concatenated to give a single first price in the sequence, then divide by the first price in the
prediction. sequence. This form of normalization is called relative change.
Normalization means that the deep learning network will look
III. DESCRIPTION OF DATASET at the degree to which the prices changed over time, rather
than the raw prices themselves, allowing it to generalize to a
The S&P 500 index (formerly Standard & Poor's 500 wider range of input values.
Index) is a market-capitalization-weighted index of the 500
largest U.S. publicly traded companies by market value. The
Output
Layer

Figure 3. Diagram of the best performing proposed multiple pipeline model with CNNs followed by Bidirectional LSTMs.

IV. PROPOSED MULTIPLE PIPELINE CNN AND BI-DIRECTIONAL sequence of data, and uses concatenation to merge the two
LSTM MODEL sequence outputs. Dropout layer is used to prevent overfitting
in the model. The use of LSTM also prevents the problem of
We constructed various deep learning architectures and vanishing gradients.
compared their performance to find the best performing
model. We also implement a Support Vector Machine Each pipeline takes the entire sequence of 50 prices as
Regressor model for comparison. A SVM Regressor model input, and the CNN layers use 128 kernel windows of size 5 or
was created using Scikit-learn [20] library in python. We 9, to return another sequence of smaller size.
perform grid search to find the best parameters for C and The output from the CNN group is fed into the Bi-
gamma for a model using a grid of both radial basis function directional LSTM layer group which:
kernel and linear kernel.
1) learns using 200 pairs of forward and backward LSTM
Then, we test several single pipeline deep-learning models units.
using using CNN and Bi-directional LSTM. We find the best
single pipeline model, configured as group of 1-dimensional 2) gives an output sequence of length 400
Convolutional layers, followed by two Bi-directional LSTM
layers with dropout. The CNN layers use ReLU as the 3) drops 50% (dropout) of the weights to generalize the model.
activation function and is followed by a Max-pooling layer The two dimensional output of this group is then fed to a
with pool size of 2. Fig. 2 diagram shows basic structure of dense layer with one unit and a linear activation, which then
single pipeline CNN based Bi-directional LSTM network. We gives the output for each individual pipeline.
use this in our experiments as a baseline deep learning model
and compare its results with support vector regressor and The output of each of the three pipeline is concatenated
proposed novel multiple pipeline deep learning model. and given to the final layer consisting of a dense layer. The
final output is a numeric value of predicted price for seven
In our proposed model, we construct a deep learning days into the future. As a typical regression problem, we have
architecture consisting of three pipelines of separate layers, used mean squared error (MSE) as the loss function to
see Fig. 3. Each pipeline consists of initial group of three 1- measure the error in the predicted versus actual data. Adadelta
dimensional CNNs with a ReLU activation followed by Max- is the gradient descent optimization algorithm for correcting
pooling. The CNNs are used so as to extract the higher level the weights in the model.
features before passing data to the Bi-directional LSTM neural
layers. The group of CNNs are followed by a group of two Bi-
directional LSTM layers with dropouts. Each Bi-directional
LSTM unit learns from both the forward and backward
TABLE I TABLE III
RESULTS FROM FOUR BEST PERFORMING SUPPORT VECTOR MACHINE RESULTS FROM FOUR BEST PERFORMING PROPOSED MULTIPLE PIPELINES
REGRESSOR MODELS ON S&P 500 DATASET DEEP LEARNING MODELS WITH CNN AND BI-DIRECTIONAL LSTM ON S&P
500 DATASET
Kernel C Gamma Mean Test Mean Train
Score Score Kernel LSTM Units Mean Test Score Mean Train
Score
rbf 0.059636233 0.212095088 0.001619057 0.001367225
9 200 0.000281317 0.000204378
rbf 0.040949150 0.372759372 0.001652736 0.001349042
5 400 0.000354261 0.000284712
rbf 0.126485521 0.120679264 0.001666914 0.001444660
5 200 0.000359209 0.000234630
rbf 0.086851137 0.212095088 0.001672472 0.001433185
9 400 0.000420501 0.000345472
Note: rbf is radial basis function. Mean Test Score and Mean Train Score are
computed from Mean Squared Error. Lower values are better. Note: Mean Test Score and Mean Train Score are computed from Mean
Squared Error. Lower values are better.

V. EXPERIMENTAL ENVIRONMENT
validation.
We have developed the model using the python
programming language. We used the following libraries for • Matplotlib [21] was used for plotting the graphs for the
various functionalities: actual time-series as well as the predicted trends.
• Keras [17] is a popular open-source library that enables • Pandas [22] was used for reading value from csv files as
researchers and software engineers to rapidly define and train DataFrames.
many deep learning models in a short amount of time. It is
used for creating, and training neural networks. It provides a • Numpy [23] used to perform matrix operations like flip,
simple interface to existing libraries like Tensorflow [18], reshape, and create random matrices.
which allows us to use GPUs for faster training and prediction. All training and prediction have been performed on several
With Keras, we can achieve fast experimentation which is the servers with multiple NVIDIA GeForce GTX 1080 Ti GPUs
key component to gain insightful feedback and garner for parallel processing, each GPU has 12GB of VRAM. The
improvements in accuracy of deep learning models for stock servers use Ubuntu 16.10 as the operating system. Main
prediction. Note that in Keras version 2.0, the Keras team has processor used is a collection of 4 Intel Xeon E5-2630 CPUs
removed metrics such as recall, precision, and fmeasure to @ 2.20 Ghz, with 10 cores and 256 GB RAM. Each server
implicitly promote the use of accuracy as the main metric for uses CUDA toolkit v9.0.176 for utilizing the Nvidia GPUs.
CNNs that are trained on balanced datasets [19]. We used two such servers with a total of 11 GPUs, allowing to
• Tensorflow [18] is the backend for Keras, and facilitates train multiple models simultaneously and consequently
the processing of low-level operations such as tensor products drawing on the benefits from parallel architecture the
and convolutions. It is developed and maintained by Google. proposed multiple pipeline model.

• Scikit-learn [20] was used for performing the grid search VI. RESEARCH RESULTS AND ANALYSIS
of the model to find the best parameters using 5-fold Cross- We have used the grid search to perform five-fold cross

TABLE II
RESULTS FROM FOUR BEST PERFORMING SINGLE PIPELINE DEEP LEARNING
MODELS WITH CNN AND BI-DIRECTIONAL LSTM ON S&P 500 DATASET

Kernel LSTM Units Mean Test Score Mean Train


Score

9 200 0.000345951 0.000272510

5 200 0.000429588 0.000324599

9 400 0.000431591 0.000333719

5 400 0.000555343 0.000466796 Figure 4: Support Vector Regressor predictions on S&P 500 Dataset. X axis
shows the time segment for the index and prediction window. Y axis shows the
closing price. Continuous plot is the actual closing price, predictions are shown
Note: Mean Test Score and Mean Train Score are computed from Mean
as disjoint plots for next seven days. Data comes from years 2015 and 2016.
Squared Error. Lower values are better.
optimal parameters. Kernel sizes of 5 or 9 with Bi-directional
LSTM units as 200 or 400 resulted in four best performing
models. All mean squared error results given by the deep
learning model were in the order of 10-4.
C. Proposed Multiple Pipeline Models Results
Proposed multiple pipeline deep learning model results are
shown in Table III. CNN’s kernel sizes of 5 or 9 with Bi-
directional LSTM units as 200 or 400 resulted in four best
performing models. The kernel size of 9 for the CNN layer
was found to perform the best in combination with the Bi-
directional LSTM layers having 200 units. Mean squared error
Figure 5: Single Pipeline Deep Learning Model predictions with CNN and Bi- results are in the order of 10-4.
Directional LSTM on S&P 500 Dataset. X axis shows the time segment for the
index and prediction window. Y axis shows the closing price. Continuous plot D. Comparison of Results Across Models
is the actual closing price, predictions are shown as disjoint plots for next seven
days window. Data comes from years 2015 and 2016. Multiple and single pipeline deep learning models
outperform Support vector machine regressor models by a
factor higher than 4.7 for MSE value. Best performing
validation for each of support vector regressor, single pipeline
multiple pipeline model outperforms single pipeline model by
and multiple pipeline models to find the top four performing
around 9%. This validates our idea that combining the three
models and their parameter values in each of the three
pipelines results in improved performance, thereby the fusion
aforementioned categories.
is non-redundant. Since, all of the individual pipelines in the
A. Support Vector Machine Regressor Models Results multiple pipelines model can be evaluated in parallel; there is
Results from the four best performing Support Vector negligible difference in training time in comparison to single
Regressor models are shown in Table I. Five-fold cross pipeline model.
validation was performed along with grid search to select best Best performing model for each of multiple pipelines,
kernel, C and gamma values. Mean squared error is used as single pipeline, and SVM regressor methodologies is used for
the performance metric. Lower value for mean squared error is constructing prediction graphs in Figs. 4, 5, 6. X axis shows
considered better. From the list of rbf and linear kernels, the the time segment for the index and prediction window. Y axis
grid search returned rbf kernels as the best performing models. shows the closing price. Continuous plot is the actual closing
This was an expected result, as radial basis function kernels price, predictions are shown as disjoint plots for next seven
generally perform better than linear kernels on non-linear data days window. Predictions are based on data from past 50 days.
[24]. The mean squared error observed for SVM regressor is Data comes from years 2015 and 2016.
in the order of 10-3. From the results it is observed that there is Predictions from the multiple and single pipelines model
low variance across best four models. Figs. 5 and 6 follow the actual prices more closely than the
support vector machine regressor model in Fig. 4. The
B. Single Pipeline Models Results predicted trends in Figs. 5 and 6 give much better stock
Single pipeline deep learning model results are shown in market direction (up or down) when compared to Fig. 4. In
Table II. Various parameter values for CNN’s kernel sizes and general deep learning models significantly outperform the
number of Bi-directional LSTM units were searched to obtain support vector machine regressor model, both in terms of
MSE and prediction trend given in Figs. 4, 5, 6.
VII. CONCLUSION AND FUTURE WORK
In this paper we proposed a new deep learning model that
combines multiple pipelines of CNN for feature extraction
from data and Bi-directional LSTM in analyzing temporal
data. We observed that CNN layers when combined with Bi-
directional LSTM show improved prediction performance
from the temporal sequence as compared to the traditional
SVM regressor. Furthermore, multiple pipelines of deep layers
that concatenate the best of the results from three individual
pipelines performs better than single pipeline model. This
demonstrates the applicability of deep learning models on
predicting both the values and trends in a time-series data. We
Figure 6: Multiple Pipeline Deep Learning Model predictions with CNN and also presented multiple variations of our model to show how
Bi-Directional LSTM on S&P 500 Dataset. X axis shows the time segment for we have increased accuracy and minimized the effects of
the index and prediction window. Y axis shows the closing price. Continuous
plot is the actual closing price, predictions are shown as disjoint plots for next overfitting.
seven days window. Data comes from years 2015 and 2016.
Future research would apply deep learning toward [12] CM. Langkvist, L. Karlsson, and A. Loutfi, "A review of unsupervised
combining model for predicting price and trends with feature learning and deep learning for time-series modeling," Pattern
Recognition Lett., vol. 42, pp. 11-24, Jun. 2014.
sentiment analysis on news data to get even better stock price
[13] W. Zhang, "Shift-invariant pattern recognition neural network and its
index prediction. optical architecture," Proc. annual conf. Japan Society of Applied
Physics,1988.
REFERENCES
[14] A.O. Russakovsky et al., "ImageNet large scale visual recognition
[1] J. Hall, G. Mani, and D. Barr, "Applying Computational Intelligence to challenge," Int. J. Comput. Vision, vol. 115, pp. 211–252, Dec. 2015.
the Investment Process", in Proc. of 1996 CIFER: Computational [15] A. Yenter and A. Verma, "Deep CNN-LSTM with Combined Kernels
Intelligence in Financial Engineering. from Multiple Branches for IMDb Review Sentiment Analysis," 8th
[2] Yahoo Finance, S&P500 stock data [Online]. Available: IEEE Annu. Ubiquitous Computing, Electron. & Mobile Commun.
https://finance.yahoo.com/quote/%5EGSPC/history?p=%5EGSPC. Conf., (UEMCON), doi: 10.1109/UEMCON.2017.8249013, Oct. 2017.
[Accessed 29 Nov. 2018]. [16] M. M. Rounaghi and F. N. Zadeh, "Investigation of market efficiency
[3] J. M. Marynowski et al., "Automated Trading System in an Electronic and Financial Stability between S&P 500 and London Stock Exchange:
Trading Exchange", U.S. Patent 7 177 833B1, Feb 13, 2007. Monthly and yearly Forecasting of Time Series Stock Returns using
[4] P. Meesad and R.I Rasel, "Predicting Stock Market Price Using Support ARMA model," Physica A: Statistical Mechanics and Its Applicat., vol.
Vector Regression," in Proc. of Int. Conf. Informatics, Electron. and 456, pp. 10-21, Aug. 2016..
Vision, (ICIEV), 2013. doi: 10.1109/ICIEV.2013.6572570. [17] Keras: The Python Deep Learning library [Online]. Available:
[5] G. Zhang, B.E. Patuwo, and M.Y. Hu, "Forecasting with artificial neural https://keras.io. [Accessed 29 Nov. 2018].
networks:: The state of the art," Int. J. Forecasting, vol. 14, pp. 35-62, [18] TensorFlow: An open source machine learning framework for everyone
Mar. 1998. [Online]. Available: https://www.tensorflow.org/. [Accessed 29 Nov.
[6] E. Guresen, G. Kayakutlu, and T.U Daim, "Using artificial neural 2018].
network models in stock market index prediction," Expert Systems with [19] Keras Team, Keras 2.0 release notes [Online]. Available:
Applications, vol. 38, pp. 10389-10397, Aug. 2011. https://github.com/keras-team/keras/wiki/Keras-2.0-release-notes.
[7] N. V. Bhattacharjee and E. W. Tollner, "Improving management of [Accessed 29 Nov. 2018].
windrow composting systems by modeling runoff water quality [20] Scikit-Learn: Machine Learning in Python [Online]. Available:
dynamics using recurrent neural network," Ecological Modelling, vol. https://scikit-learn.org/stable/. [Accessed 29 Nov. 2018].
339, pp. 68-76, Nov. 2016. [21] The Matplotlib development team (2018, Nov. 28). matplotlib Version
[8] M. Herman and B. Schrauwen, "Training and Analyzing Deep Recurrent 3.0.2 [Online]. Available: https://matplotlib.org/. [Accessed 29 Nov.
Neural Networks," Advances in Neural Information Processing Systems 2018].
26, (NIPS 2013). [22] T. Mester (2018, Jul. 10). Pandas Tutorial 1: Pandas Basics [Online].
[9] S. Hochreiter and J. Schmidhuber, "Long Short-Term Memory," Neural Available: https://data36.com/pandas-tutorial-1-basics-reading-data-
Computation, vol. 9, pp. 1735-1780, 1997. files-dataframes-data-selection/. [Accessed 29 Nov. 2018].
[10] W. Bao, J. Yue and Y. Rao, "A deep learning framework for financial [23] NumPy [Online], (2018). Available: http://www.numpy.org/. [Accessed
time," PLoS ONE, vol. 12, doi: 10.1371/journal.pone.0180944, Jul. 29 Nov. 2018].
2017. [24] C.W. Hsu, CC. Chang, and C.J. Lin, "A practical Guide to Support
[11] Y.M. Schuster and K.K. Paliwal, "Bidirectional Recurrent Neural Vector Classification," Dept. Comp. Sci., Nat. Taiwan Univ., Taipei,
Networks," IEEE Trans. Signal Process., vol. 45, pp. 2673-2681, Nov. Apr. 2010
1997.

You might also like