Documente Academic
Documente Profesional
Documente Cultură
DOI 10.1007/s11222-013-9375-7
Abstract We introduce a novel predictive statistical modeling technique called Hybrid Radial Basis Function Neural
Networks (HRBF-NN) as a forecaster. HRBF-NN is a flexible forecasting technique that integrates regression trees,
ridge regression, with radial basis function (RBF) neural
networks (NN). We develop a new computational procedure
using model selection based on information-theoretic principles as the fitness function using the genetic algorithm (GA)
to carry out subset selection of best predictors. Due to the
dynamic and chaotic nature of the underlying stock market process, as is well known, the task of generating economically useful stock market forecasts is difficult, if not
impossible. HRBF-NN is well suited for modeling complex
non-linear relationships and dependencies between the stock
indices. We propose HRBF-NN as our forecaster and a predictive modeling tool to study the daily movements of stock
indices. We show numerical examples to determine a predictive relationship between the Istanbul Stock Exchange
National 100 Index (ISE100) and seven other international
stock market indices. We select the best subset of predictors
by minimizing the information complexity (ICOMP) criterion as the fitness function within the GA. Using the best
subset of variables we construct out-of-sample forecasts for
O. Akbilgic () M.E. Balaban
Istanbul University School of Business Administration, Istanbul,
Turkey
e-mail: oguzakbilgic@gmail.com
M.E. Balaban
e-mail: balaban@istanbul.edu.tr
H. Bozdogan
Statistics, Operations, and Management Science, and Center
for Intelligent Systems and Machine Learning (CISML),
The University of Tennessee, Knoxville 37996, USA
e-mail: bozdogan@utk.edu
the ISE100 index to determine the daily directional movements. Our results obtained demonstrate the utility and the
flexibility of HRBF-NN as a clever predictive modeling tool
for highly dependent and nonlinear data.
Keywords Forecasting Stock markets Neural networks
Variable selection Radial basis functions
1 Introduction
We introduce a novel predictive statistical modeling approach called Hybrid Radial Basis Function Neural Networks (HRBF-NN) as a forecaster. HRBF-NN is a flexible
forecasting technique that integrates regression trees, and
ridge regression, with radial basis function (RBF) neural
networks (NN). We develop a new computational technique
based on information complexity (ICOMP) criterion (Bozdogan 1994, 2000, 2004) as the fitness function within the
genetic algorithm (GA) to carry out subset selection of best
predictors.
New generation research argue that one needs to look at
financial market data on stock indices from an evolutionary
or more from an adaptive biological points of view rather
than looking at such problems from physical systems points
of view. Traditional modeling techniques of the financial
markets such as random walk models (i.e., the Brownian
motion ideas), and rigid stochastic processes do not adapt
and land themselves to the ever changing dynamics of the
financial data (Lo and MacKinlay 1988). They are inflexible
and do not capture some of the behavioral and psychological
issues that the financial portfolio managers and the investors
face.
It is because of this and due to the dynamic and chaotic
nature of the underlying stock market process, as is well
Stat Comput
one should take into consideration the effect of this dependency in order to make accurate forecasts. This interdependency has especially large impact on stock market indices in
developing countries, such as Turkey.
In reviewing the literature, we note that there are several
studies which show the influence of international markets on
the Istanbul Stock Exchange (ISE) and predict the direction
of movements in the ISE100 index. For example, Korkmaz
et al. (2011) in their causality study, showed that ISE100 index is affected by the US markets. In another study, Ozun
(2007) showed the influence of volatility of stock markets
in developing countries along with Turkish and Brazilian
stock markets. In his study, Ozun (2007) also showed that
the US markets have influence in the positive direction,
i.e., when the US markets go up, the ISE100 index goes
up. Further, Vuran (2000) showed that the ISE100 index
is co-integrated with stock markets of the United Kingdom
(FTSE), Brazil (BOVESPA), and Germany (DAX). However, in their study, Boyacioglu and Avci (2010) used the
BOVESPA, Dow Jones Industrials (DJI), and DAX indices
along with other macro- and micro-economic variables as
predictors to forecast the ISE100 index. Cinko and Avci
(2007) compared artificial neural networks and regression
models in forecasting the daily values of the ISE100. Their
results show that neural networks models perform better
than the classical regression model. In their study, they used
only the lagged series of ISE100 as explanatory variables.
On the other hand, Ozdemir et al. (2011) used the MSCI
Emerging Markets (EM), MSCI European (EU), and S&P
500 (SP) indices along with other macroeconomic indicators to forecast the direction of movement of the ISE100 index.
Further, we also note that Radial Basis Function (RBF)
Neural Networks (NN) (RBF-NN) have been used as an alternative method in forecasting problems in stock markets.
The RBF-NN (Broomhead and Lowe 1988) model is a
special type of feed-forward neural network with one input, one hidden, and one linear output layer. In RBF-NN
the number of parameters is fewer than Multilayer Neural
Networks (ML-NN) because inputs are directly connected
to the hidden layer without using weights (Haykin 1999).
Center, width, and weight parameters of RBF-NN are the adjustable parameters. Moreover, center and width parameters
belong to the hidden layer of RBF-NN. On the other hand,
the weight parameters are the connection weights between
the hidden and output layers. Traditional RBF-NN learning
algorithms are used to determine the best parameters using
iterative techniques such as the gradient descent, and forward selection procedures.
There are several methods proposed to automatically
construct an RBF-NN model. In their study, Sun et al. (2005)
combined optimal partition algorithm with RBF-NN in order to determine the center and width parameters. Further,
Stat Comput
y = f (w, x)
=
m
wj hj (x) = w1 h1 + w2 h2 + + wm hm ,
(1)
j =1
1 + exp(
1
p
(xk cj k )2
)
k=1
rj2k
(3)
(4)
(xk cj k )2
)
rj2k
(5)
In short, the RBF-NN introduces a mapping or transformation of the n-dimensional inputs non-linearly to an mdimensional space and then estimates a model using linear
regression. The non-linear transformation is achieved using
m basis functions, each characterized by their centers cj in
the (original) input space and a width or radius vector rj ,
j {1, 2, . . . , m} (Orr 2000). Poggio and Girosi (1990) has
shown that RBF-NN possess the property of best approximation.
3.2 Estimation of weights: ridge regression
p
2
f (xi ) yi
i=1
(6)
Stat Comput
and is given by
1
w = H H
H y,
(7)
p
f (xi ) yi
i=1
m
wj2 = + w w
(8)
i=1
which is minimized to find a weight vector that is more robust to noise in the training set. The optimal weight vector
for global ridge regression is given by
1
w = H H + Im H y,
(9)
ms 2
w LS w LS
(10)
which is a data adaptive approach. Here, m = k is the number of predictors not including the intercept term, n is the
number of observations, s 2 is the estimated error variance
using k predictors so that
s2 =
1
(y H w LS ) (y H w LS )
(n k + 1)
(11)
1
H y.
w LS = H H
(12)
max(xik ) min(xik )
iS
2 iS
1
ck = max(xik ) + min(xik )
iS
2 iS
sk =
(13)
(14)
(15)
(16)
1
yi ,
pL
(17)
1
yi ,
pR
(18)
iSL
yR =
iSR
iSR
Stat Comput
(20)
The scalar has the same value for all nodes (Kubat 1998),
and
is1another parameter of the method. One can use =
2K where K is the Kubats parameter (Kubat 1998;
Orr 2000).
One of the general forms of ICOMP is an approximation to the sum of two Kullback-Leibler (KL) (Kullback and
Leibler 1951) distances. For a general multivariate normal
linear or non-linear structural model this general form of
ICOMP is given by
ICOMP(IFIM) = 2 log L( ) + 2C1 F 1 ( ) ,
(21)
s
tr(F ( ))
C1 F 1 ( ) = log L
2
s
1
log F 1 ( ) ,
2
(22)
2 (H H )1
0
2
1
w,
Cov
= F =
(23)
2 4 ,
0
4
and where
2 =
)
(y H w
) (y H w
n
(24)
ICOMP(IFIM) = n log(2) + n log L 2 + n
+ 2C1 F 1 ( )
(25)
1
tr 2 (H H )1 + 24
C1 F
= (m + 1) log
m+1
4
2
1
2
1
+ log
log H H
2
4
(26)
)Misspec ,
= 2 log L( ) + 2C1 Cov(
(27)
Stat Comput
where
F 1 .
)Misspec = F 1 R
Cov(
(28)
2
=
R
(29)
(nq)(Kt1) ,
Sk
(H 1 2
)
3
4
4
where
12 , . . . ,
n2 ),
D 2 = diag(
H is (n q) matrix of regressors or model matrix,
Sk is the estimated residual skewness,
Kt the kurtosis, and 1 is a (n 1) vector of ones.
ICOMP(IFIM)Misspec = n log(2) + n log 2 + n
)Misspec .
+ 2C1 Cov(
pling (Liang and Wong 2001). There are several applications of the GA in a variety of fields including econometrics (Routledge 1999), finance (Neely et al. 1997), and image processing (Bhandarkar et al. 1994). Additionally, successful application of the GA on variable selection has also
been reported in the literature (Bozdogan and Howe 2012;
Howe and Bozdogan 2010).
Recall that the regularized regression tree and RBF networks model given in (1), the GA is used to find the best or
nearly best subset of predictors from the data.
The summary of the implementation of the GA is as follows.
(30)
(31)
n
1
ICOMP(i)
N
i=1
(32)
Stat Comput
ICOMP(i)
ICOMP
(33)
Variable explanation
ISE100
SP
DAX
FTSE
NIK
BVSP
EU
EM
Stat Comput
Table 2 Variable selection with the genetic algorithm
Mut. pr.
Xover pr.
RBF
ICOMP
Variable subset
0.65
0.010
Gauss
2773.4
0111111000101101
0.85
0.010
Gauss
2773.1
1111111000101111
0.65
0.010
IMQ
2772.9
0110111111101010
0.65
0.005
IMQ
2771.4
0010010111101000
0.65
0.010
Cauchy
2771.4
0110011011101000
0.60
0.050
Gauss
2770.8
0010000111101011
0.50
0.100
Gauss
2770.7
0110011111101000
0.80
0.005
Gauss
2770.6
0110010111101011
0.75
0.010
Gauss
2770.6
0110010111101011
0.60
0.100
Gauss
2769.6
0010111011101001
are the one-day lags of the indices listed in Table 1 while the
other eight predictors are the two days lags of the same indices. As previously stated, the ICOMP(IFIM)Misspec criterion is used to score the subset models as the fitness function
within the genetic algorithm (GA). We used a population
size of N = 25, which is larger than the number of predictors, 16. Although, in our experiments, there was generally
no improvement on the fitness function after 18th generation, we let the GA run for 30 iterations. We ran genetic algorithm for different values of crossover and mutation probabilities attempting to minimize the ICOMP score. In our
experimental study, we set the crossover and mutation probabilities from the sets {0.50, 0.55, 0.60, 0.65, 0.70, 0.75,
0.80, 0.85, 0.90, 0.90} and {0.005, 0.01,0.05, 0.1}, respectively.
During our experimental study, we ran the GA for variable selection for different mutation and crossover probabilities, and the RBF combinations, 160 times in total. We
ran the GA for four of the RBFs, Gaussian, Cauchy, Multi
Quadratic, and Inverse Multi Quadratic. The results for best
ten out of 160 runs for different parameter combinations are
listed in Table 2. During each run of the GA elitism rule
has been applied in order to keep the best subset in each
generation of the GA found so farthis ensures monotonic
improvement of the fitness function.
Table 2 shows that, the fitness function, ICOMP, is minimized at the GA parameters: Crossover Probability = 0.65,
Mutation Probability = 0.01, and the Gaussian RBF. Therefore, we chose the Gaussian RBF as the best fitting RBF for
forecasting the ISE100 index. The best fitting model with
our approach is given by the first variable subset in Table 2.
The predictors of best subset, represented as binary code,
correspond the variable subset, {SP1, DAX1, FTSE1, NIK1,
BVSP1, EU1, DAX2, NIK2, BVSP2, EM2}. Note that the
ISE1 and ISE2 variables are not in the best subset with minimum ICOMP value. This means that the ISE100 is strongly
affected by the lags of the international markets, not by its
own lags. This is an interesting as well as an important result
Stat Comput
Fig. 1 Movement of invested
$100 in 200 days
Forecast terms
No buy-sell
Day
Beginning (Dollar)
100
100
Period 1
117.83
91.27
Period 2
127.24
90.62
Period 3
145.85
104.51
Period 4
153.34
99.72
Period 5
171.96
114.43
Period 6
185.07
131.00
Period 7
162.70
113.36
Period 8
176.53
112.08
Period 9
186.17
108.94
Period 10
202.06
105.52
tion costs into consideration, our investment would have become $116.90. If we had just bought and held the index, the
investment would have declined to $91.27.
6 Conclusions
In this paper, we studied a variable selection and forecasting
problem in stock markets by focusing on the ISE100 index.
We have identified a model for the effects of international
stock markets on the ISE100 index. We carried out a variable
subset selection using the GA along with ICOMP criterion
to determine which indices have important effects on the direction of the movement of the ISE100 index. Our variable
subset selection results selected the first and second lags of
NIK, DAX, and BVSP and only first lags of SP, FTSE, and
EU, and only second lag of EM indices as explanatory variables. It is interesting to see that none of the lags of ISE100
were selected as explanatory variables. Our results suggest
that the ISE100 index does not require its lag variables to
build a predictive model.
ISE100
Forecast
Decision
HRBF-NN
No buy-sell
Sell
100.0000
97.2299
Buy
97.5386
94.8367
Sell
97.5386
87.8112
Buy
107.8638
97.1066
Sell
107.8638
96.0809
Keep
107.8638
99.1060
Buy
108.5335
99.7213
Sell
108.5335
95.1962
Keep
108.5335
94.8693
10
Buy
110.8596
96.9026
11
Sell
110.8596
90.0334
12
Keep
110.8596
89.3095
13
Buy
112.5537
90.6742
85.6230
14
Sell
112.5537
15
Buy
117.5910
89.4550
16
Keep
119.3674
90.8064
17
Keep
121.7044
92.5842
18
Keep
118.6782
90.2821
19
Keep
117.8346
89.6404
20
Sell
117.8346
91.2655
Stat Comput
sumption and consider a more general distributional assumption on the random noise such as the Power Exponential (PE) distribution. The PE distribution includes the
Gaussian, Laplace, and other distributions as a subfamily.
Our results will be reported in a subsequent paper elsewhere.
All our computations has been carried out using newly
developed scripts in MATLAB . Since there is still ongoing
research being done by the first two authors in HRBF-NN
these scripts presently are not freely available.
Acknowledgements This work was supported by Scientific Research Projects Coordination Unit of Istanbul University under project
number 17708. We further acknowledge the valuable comments of the
three anonymous referees and the Associate Editor which resulted to a
much improved paper.
References
Akaike, H.: Information theory and an extension of the maximum likelihood principle. In: Petrox, B.N., Csaki, F. (eds.) Second International Symposium on Information Theory, pp. 267281. Akad.
Kiad, Budapest (1973)
Akbilgic, O., Bozdogan, H.: Predictive subset selection using regression trees and RBF neural networks hybridized with the genetic
algorithm. Eur. J. Pure Appl. Math. 4(4), 467485 (2011)
Bhandarkar, S., Zhang, Y., Potter, W.: Edge detection technique using genetic algorithm-based optimization. Pattern Recognit. 27,
11591180 (1994)
Bishop, C.: Improving the generalization properties of radial basis
function neural networks. Neural Comput. 3(4), 579588 (1991)
Boyacioglu, M., Avci, D.: An adaptive network-based fuzzy inference
systems (ANFIS) for prediction of stock market return: the case
of Istanbul stock exchange. In: Expert Systems with Applications,
vol. 37, pp. 79027912 (2010)
Bozdogan, H.: ICOMP: a new model-selection criteria. In: Bock, H.H.
(ed.) Classification and Related Methods of Data Analysis. NorthHolland, Amsterdam (1988)
Bozdogan, H.: Mixture-model cluster analysis using a new informational complexity and model selection criteria. In: Proceedings of
the First US/Japan Conference on the Frontiers of Statistical Modeling: An Informational Approach. Multivariate Statistical Modeling, vol. 2, pp. 69113. Kluwer Academic, Norwell (1994)
Bozdogan, H.: Akaikes information criterion and recent developments
in informational complexity. J. Math. Psychol. 44, 6291 (2000)
Bozdogan, H.: Intelligent statistical data mining with information complexity and genetic algorithms. In: Bozdogan, H. (ed.) Statistical
Data Mining and Knowledge Discovery, pp. 1556. Chapman &
Hall, London (2004)
Bozdogan, H., Howe, J.A.: Misspecified multivariate regression models using the genetic algorithm and information complexity as the
fitness function. Eur. J. Pure Appl. Math. 5(2), 211249 (2012)
Breiman, L., Freidman, J., Stone, J.C., Olsen, R.A.: Classification and
Regression Trees. Chapman & Hall, London (1984)
Broomhead, D.S., Lowe, D.: Multi-variable functional interpolation
and adaptive networks. Complex Syst. 11, 321355 (1988)
Burns, P.: A genetic algorithm for robust regression estimation. Technical report from Statistical Sciences, Inc. (1992)
Cinko, M., Avci, E.: A comparison of neural network and linear regression forecast of the ISE100 index. neri 7(28), 301307 (2007)
De Jong, K.A.: An analysis of the behavior of a class of genetic adaptive systems. Ph.D. Dissertation, University of Michigan (1975)
De Jong, K.A., Spears, W.M.: Using genetic algorithms to solve NPcomplete problems. In: Schaffer, J.D. (ed.) Third Conference on
Genetic Algorithms, pp. 124132. Morgan Kaufmann, San Mateo
(1989)
Eiben, A.E., Smith, J.E.: Introduction to Evolutionary Computing.
Springer, Berlin (2010)
Fouskakis, D., Draper, D.: Stochastic optimization: a review. Int. Stat.
Rev. 70(2), 315349 (2002)
Hamada, M., Martz, H., Reese, C., Wilson, A.: Finding near-optimal
Bayesian experimental designs via genetic algorithms. Am. Stat.
55(3), 175181 (2001)
Haykin, S.: Neural Networks: A Comprehensive Foundation. Prentice
Hall, New Jersey (1999)
Hoerl, A.E., Kennard, R.W., Baldwin, K.F.: Ridge regression: some
simulations. Commun. Stat. 4, 105123 (1975)
Howe, J.A., Bozdogan, H.: Predictive subset VAR modeling using the
genetic algorithm and information complexity. Eur. J. Pure Appl.
Math. 3(3), 382405 (2010)
Howlett, R.J., Jain, L.C.: Radial Basis Function Networks 1: Recent
Developments in Theory and Applications. Physica-Verlag, New
York (2001)
Korkmaz, T., Cevik, E., Birkan, E., Ozatac, N.: Causality in mean and
variance between ISE100 and S&P 500: Turkcell case. Afr. J. Bus.
Manag. 5(5), 16731683 (2011)
Kubat, M.: Decision trees can initialize radial basis function networks.
IEEE Trans. Neural Netw. 9(5), 813821 (1998)
Kullback, A., Leibler, R.: On information and sufficiency. Ann. Math.
Stat. 22, 7986 (1951)
Liang, F., Wong, W.: Real-parameter evolutionary Monte Carlo with
applications to Bayesian mixture models. J. Am. Stat. Assoc.
96(454), 653666 (2001)
Lin, C.-T., Lee, C.S.G.: Neural Fuzzy Systems: A Neuro-Fuzzy Synergism to Intelligent Systems. Prentice Hall, New York (1996)
Liu, Z., Bozdogan, H.: Improving the performance of radial basis function classification using information criteria. In: Bozdogan, H.
(ed.) Statistical Data Mining and Knowledge Discovery, pp. 193
216. Chapman & Hall, London (2004)
Lo, A.W., MacKinlay, C.: Stock market prices do not follow random
walks: evidence from a simple specification test. Rev. Financ.
Stud. 1, 4166 (1988)
Meng, K., Dong, Z.Y., Wong, K.P.: Self-adaptive radial basis function
neural networks for short-term electricity price forecasting. IEE
Proc., Gener. Transm. Distrib. 3(4), 325335 (2008)
Neely, C., Weller, P., Dittmar, R.: Is technical analysis in the foreign
exchange market profitable? A genetic programming approach.
J. Financ. Quant. Anal. 32(4), 405426 (1997)
Orr, M.: Combining regression trees and RBFs. Int. J. Neural Syst.
10(6), 453465 (2000)
Ozdemir, A.K., Tolun, S., Demirci, E.: Endeks getirisi yonunun ikili siniflandirma yontemiyle tahmin edilmesi: IMKB100 endeksi
ornegi. Nigde Univ. IIBF Derg. 4(2), 4559 (2011)
Ozun, A.: Are the reactions of emerging equity markets to the volatility
in advanced markets similar? Comparative evidence from Brazil
and Turkey. Int. Res. J. Finance Econ. 9, 220230 (2007)
Poggio, T., Girosi, F.: Regularization algorithms for learning that are
equivalent to multilayer networks. Science 247(4945), 978982
(1990)
Rivas, V.M., Merelo, J.J., Castillo, P.A., Arenas, M.G., Castellano,
J.G.: Evolving RBF neural networks for time-series forecasting
with EvRBF. Inf. Sci. 1655(5354), 207220 (2004)
Routledge, B.: Adaptive learning in financial markets. Rev. Financ.
Stud. 12(5), 11651202 (1999). Oxford University Press
Srinivas, M., Patnaik, L.M.: Adaptive probabilities of crossover and
mutation in genetic algorithms. IEEE Trans. Syst. Man Cybern.
24(4), 656667 (1994)
Stat Comput
Sun, Y.F., Liang, Y.C., Zhang, W.L., Lee, H.P., Lin, W.Z., Cao, L.J.:
Optimal partition algorithm of the RBF neural network and its
application to financial time series forecasting. Neural Comput.
Appl. 14, 3544 (2005)
Tikhonov, A.N., Arsenin, V.Y.: Solutions of Ill-Posed Problems. Wiley,
New York (1977)
Vuran, B.: The determination of long-run relationship between ISE100
and international equity indices using cointegration analysis. Istanb. Univ. J. Sch. Bus. Adm. 39(1), 154168 (2000)