Professional Documents
Culture Documents
* University
of Hawaii
** University
Abstract
Neural networks are advocated as a replacement for statistical forecasting methods. In this paper we review the literature comparing neural networks and classical forecasting methods, particularly in causal forecasting, time series forecasting, and judgmental forecasting. We provide not only an overview and evaluation of the literature but akso summarize several studies we have performed which address the typical criticisms of work in this area. Overall, the empirical studies find neural networks at least as good as their classical counterparts.
K y Words: Neural Networks, Regression, Forecasting, e Judgmental Forecasting, Time Series. Introduction
Over the last few decades, there has been much research directed at understanding and predicting the future. This research has led to many developments in forecasting methods - including many methodological advances based on statistical techniques. Currently, there is a new challenger for these methodologies - neural networks. Neural networks (NN) have been widely touted as solving many forecasting problems. For example, they are argued to be able to model easily any type of parametric or non-parametric process including automatically and optimally transforming the input data. These sorts of claims have led to much interest in neural networks. In this paper we review the literature comparing neural networks and the classical forecasting techniques. Our review will be segmented into three different applications areas: time series forecasting, regression-based causal forecasting, and regression-based judgmental forecasting. Additionally, we will note the literature comparing neural networks and other linear models such as discriminant analysis and logit models.
494
with long memory, neural network models and BoxJenkins models produce comparable results. For time series with short memory, they argued that neural networks outperform the Box-Jenkins models. When forecasting nonlinear dynamic systems with chaotic behavior, Lapedes and Farber (1987) showed that neural networks can be orders of magnitude more accurate than conventional methods. Neural networks have also been implemented for diagnostic applications in time series analysis. Sastri et al. (1990) compared the performance of backpropagation, counterpropagation, and categocy learning neural network models in identifying the auto-regressive time series models. The backpropagation network was best in identifying the correct time series model while the counterpropagation network was best in detecting changes in the model. Foster, Collopy, and Ungar (1991) had similar success in using neural networks to identify the best time series model. As part of the neural network sessions at Hawaii International Conference of System Sciences, the referees developed a list of desirable characteristics for such studies (see Exhibit 2). AU the above studies (except Foster, Collopy, and Ungar) can be faulted on one or more of the points in that list. Those studies not using Mcompetition data generally had far too few series and too few data points per series to support generalization; thus, they are only suggestive. The Sharda and Patil study (1990b) mysteriously threw away one third of the series; the latter leads us to question the generalizeability of their results. Foster, Collopy, and Ungar (1991) didnt have that problem but found neural networks to be inferior to the classical methods. In comparing forecasting models it is important to develop the model on one set of data and then test it on another. In neural network studies, the experimenters often develop many neural network models and compare them on the holdout data. Then they select the best to compare with the classical models. This may have happened in some of the above studies; that is, the models appeared partially fit on the holdout data (rather than using the holdout data just for the final test comparison). AU the concerns in this and the above prior paragraph led us to conduct the following experiment. In the &U, OConnor, and Remus (1991) research, time series forecasts based on neural networks were compared with forecasts from six statistical time series methods (including exponential smoothing and BoxJenkins) and two judgment-based methods. The experiment used all 111 time series from the Mcompetition. The classical methods were all estimated by experts in a particular technique as part of the Mcompetition.
The 111 series were each divided by Makridakis et al. into a portion of model estimation and a portion for model testing. The structure of the neural networks in this study was developed using the complete series #1 and #2; the remaining 109 neural networks used the structure found in #1 and #2 but were estimated by the first author using the same ground rules as the M-competition. That is, the first author was not supplied with the holdout data (except in the cases of series #1 and a). We then compared his neural network models forecasts for the holdout data with the forecasts made by experts in the M-competition with the classical methods. Amss all 109 series, the neural networks did significantly better or as good as statistical and human judgment methods.
495
Dutta and Shekhar (1988) used 10 factors to predict the ratings of corporate bonds. They found that the neural networks outperformed the regression models. Duliba (1991) compared neural network models with four types of regression models in predicting the financial performance of transportation companies. She found that the neural network model outperformed the random effects regression model but not the fixed effects model. As in time series comparisons, these studies also suffer from having a small number of data sets and few data points per set. Also, the holdout data may have effected model estimation and specification. To overcome these problems, Remus and Hill (1990) compared the production scheduling decisions as modeled by neural networks and regression-based decision rules. The data used was from 62 decision makers who each made 24 decisions (Remus, 1987); thus, neural network and regression models were developed and compared for each of the 62 decision makers. The model structure used was the same structure used in the comparable regression models. Neural network models performed as well but not better than those using the linear regression models. In a second study (Remus and Hill, 1991), we continued the above research and aggregated the data from all 62 decision makers to estimate a composite neural network modeL The resulting neural network model performed better than both the classical models and neural networks from the earlier study; the composite neural network also performed at least as well as classical composite models. It is important to note that the above two studies did not use holdout samples so that the procedures would parallel those used in the classical models.
analysis. Practical problems where the comparison between neural networks and disaiminant analysis has been applied include the prediction of stock price performance (Yoon and Swales, 1990), the prediction of company bankruptcy (Odom and Sharda, 1990; Raghupathi et al., 1991; Koster et al., 1990), and the assignment of ratings to bonds (Surkan and Singleton, 1990). In all of these studies, the neural network model outperformed discriminant analysis. Neural networks proved to be more robust (Odom and Sharda, 1990), were able to extract higher level features for better generalization (Raghupathi et al., 1991), and detected aspects of the data set space that were beyond the capability of the statistical model (Koster et al., 1990). These studies again suffer from many of the shortcomings listed in Exhibit 2. In particular, the number of data sets is limited and the size of the data sets is small. In many cases, the holdout was iteratively used to provide information to aid in selecting the best model; therefore, the models are partially fit on the holdout data. Thus, the studies in this area are suggestive but do not critically demonstrate the advantages of neural networks.
Summary
Across the areas of time series forecasting, regression forecasting, and judgmental forecasting, neural networks seem to perform at least as well as classical models. This is still and preliminary finding and in need of more exploration. In particular, few studies have examined the (1) data conditions or (2) implementation situations favoring neural networks. Marquez et al. (1991), as an example of the former, demonstrated that neural networks are more advantageous with low sample sizes and high levels of noise. As an example of the latter, Remus and Hill (1990 and 1991) argue for the advantages of neural networks when building intelligent systems. Also there is a need for rigorous studies of the relative robustness of neural networks and classical models. The current work by Marquez will begin to address these issues. Also, most of these studies use backpropagation; perhaps new and better models might improve neural network's performance.
References
Bell, T., G. Ribar and J. Verchio, "Neural Nets vs. Logistic Regression," presented at the University of Southern California Expert Systems Symposium, November, 1989. Duliba, K., "Contrasting Neural Nets with Regression in Predicting Performance," Proceedines of the 24th Hawaii International Conference on System Sciences, Vol. 4, 163-170
496
(1991). Dutta, S . and S . Shekhar, "Bond Rating: A Non-Conservative Application of Neural Networks," Proceedings of the 1988 I " Vol. 2, 443-450 (1988). C. Foster, B., F. Collopy and L. Ungar, "Neural Network Forecasting of Short, Noisy Time Series," presented at the ORSA TIMS National Meeting, May, 1991. Koster, A., N. Sondak and W. Bourbia, "A Business Application of Artificial Neural Network Systems," The Journal of Computer Information Svstems, Vol. XXXI, 3-10 (1990). Lapedes, A. and R. Farber, "Nonlinear Signal Prediction Using Neural Networks: Prediction and System Modelling," Los Alamos National Laboratory Report LA-UR-87-2665 (1987). Makridakis, S., A. Anderson, R. Carbone, R. Fildes, M. Hibon, R. Lewandowski, J. Newton, E. Parzen, E. and R. Winkler, "The Accuracy of Extrapolation (Time Series) Methods: Results of a Forecasting Competition," Journal of Forecasting, 1, 111153 (1982). Marquez, L., T. Hill, W. Remus and R. Worthley, "Neural Network Models as an Alternative to Regression," Proceedinas of the 24th Hawaii International Conference on Svstem Sciences, Vol. 4, 129-135 (1991). Odom, M. and R. Sharda, "A Neural Network Model for Bankruptcy Prediction," Proceedings of the 1990 UCNN, Vol. 2, 163-168 (1990). Raghupathi, W., L. Schkade and R. Bapi, "A Neural Network Application for Bankruptcy Prediction," Proceedings of the 24th Hawaii International Conference on Svstem Sciences, Vol. 4, 147-155 (1991). Remus, W., "A Study of the Impact of Graphical and Tabular Displays and Their Interaction with Environmental Complexity," Management Science, Vol. 33, 1200-1205 (1987). Remus, W. and T. Hill, "Neural Network Models of Managerial Judgment," Proceedings of the 23rd Hawaii International Conference on Svstem Sciences, Vol. 4, 340-344 (1990). Remus, W. and T. Hill, "Neural Network Models for Intelligent Support of Managerial Decision Making," University of Hawaii Working Paper, (1991). Roy, J. and J. Cosset, "Forecasting Country Risk Ratings Using a Neural Network," Proceedinas of the U r d Hawaii International Conference on Svstem Sciences, Vol. 4, 327-334 (1990). Sastri, T., J. English and G. Wang, "Neural Networks for Time Series Model Identification and Change Detection," Texas A&M (1990). University Working Paper INEN/QC/WP/03/07-90
Sharda, R. and R. Patil, "Neural Networks as Forecasting Experts: An Empirical Test," Proceedings of the 1990 UCNN Meeting, Vol. 2, 491-494 (199Oa). Sharda, R. and R. Patil, "Connectionist Approach to Time Series Prediction: An Empirical Test," Oklahoma State University Working Paper 90-26 (1990b). Surkan, A. and J. Singleton, "Neural Networks for Bond Rating Improved by Multiple Hidden Layers," Proceedings of the 1990 UCNN, Vol. 2, 157-162 (1990). Tang, Z., C. de Almeida and P. Fishwick, "Time Series Forecasting Using Neural Networks vs. Box-Jenkins Methodology," presented at the 1990 International Workshop on Neural Networks, Feb., 1990. Yoon, Y.and G. Swales, "Predicting Stock Price Performance," Proceedings of the 24th Hawaii International Conference on System Sciences, Vol. 4, 156-162 (1991).
Exhibit 1 Literature Comparing Neural Networks and Time Series Methods Study Hill et al. (1991) Sastri et al. (1990) Type of Data 111 time series 92 simulated, 1 real time Results N"s were superior to all classical models. Backpropagation was the best model indentifier. Counterpropagation detected series model changes best. N"s are comparable in performance to Automatic Box-Jenkins. With short memory, NN outperformed Box-Jenkins. N"s are more accurate than conventional methods. N"s appear to be better than Box-Jenkins for long term forecasting. N"s were inferior to classical methods.
75 time series
Lapedes and Farber chaotic (1987) series Tang et al. (1990) Foster et al. (1991)
3 time series
497
Exhibit 2 The HICSS Reviewers Summary Opinions on what Constituted an Excellent Neural Network (NN) Paper The following are the desirable attributes of a research paper on neural networks for a business application: 1 The business task modeled should be important. . 2 The literature on the task and on previous approaches to modeling . the task should be well reviewed. 3. The classical non-NN approaches to the task should be compared to the N N models. 4 Multiple sets of realistic data should be used for the evaluation. The . data should be such that we can feel certain that the comparison between Ns and the classical methods applies to a wide class of similar situations. For example, when using a forecasting task, the Ns and classical models should be compared on several different real data series.
5. The NN and classical models should be evaluated on a different set of data than that on which the N N was trained and the model estimated. 6. Statistics should be reported which not only show that 4s work but also how they compare to classical non-NN approaches. The statistics should include those normally used to evaluate the classical non-NN approaches plus any other critical measures. 7. Adequate detail should be reported on the NN implementation so othen can replicate the work. For example, the software used, values of options selected for the NN,and the epochs required should be reported.
Exhibit 3 Literature Comparing Neural Networks and Regression Analysis in Judgmental Forecasting Study Study Duliba Type of Data Results Bell
Exhibit 4 Literature Comparing Neural Networks and Other Linear Models Type of Data commercial bank failures bankruptcy data for 91 business organizations country risk rating Results
(1991)
(1988)
Remus and Hill
(1990)
(1989)
Koster et al.
(1990)
1-1,1-12-1, 1-12-6-1
Ns did as well but not better than linear decision rules. Composite Ns did better than classical decision models.
(1990)
(1991)
production scheduling
(1990)
Raghupthi et al.
bankruptcy prediction bankruptcy data on 102 companies stock price data on 98 companies
(1991)
Yoon and Swales (1990)
498