ArticlePDF Available

Integrating Metaheuristics and Artificial Neural Networks for improved Stock Price Prediction

Authors:

Abstract and Figures

Stock market price is one of the most important indicators of a country's economic growth. That's why determining the exact movements of stock market price is considerably regarded. However, complex and uncertain behaviors of stock market make exact determination impossible and hence strong forecasting models are deeply desirable for investors' financial decision making process. This study aims at evaluating the effectiveness of using technical indicators, such as simple moving average of close price, momentum close price, etc. in Turkish stock market. To capture the relationship between the technical indicators and the stock market for the period under investigation, hybrid Artificial Neural Network (ANN) models, which consist in exploiting capabilities of Harmony Search (HS) and Genetic Algorithm (GA), are used for selecting the most relevant technical indicators. In addition, this study simultaneously searches the most appropriate number of hidden neurons in hidden layer and in this respect; proposed models mitigate well-known problem of overfitting/underfitting of ANN. The comparison for each proposed model is done in four viewpoints: loss functions, return from investment analysis, buy and hold analysis, and graphical analysis. According to the statistical and financial performance of these models, HS based ANN model is found as a dominant model for stock market forecasting.
Content may be subject to copyright.
Expert Systems With Applications 44 (2016) 320–331
Contents lists available at ScienceDirect
Expert Systems With Applications
journal homepage: www.elsevier.com/locate/eswa
Integrating metaheuristics and Artificial Neural Networks for improved
stock price prediction
Mustafa Göçkena,, Mehmet Özçalıcıb,AslıBorua,Ay¸se Tu˘
gba Dosdo˘
gruc
aAdana Science and Technology University, Industrial Engineering Department, Ye¸siloba Yerle ¸skesi, 01180, Adana, Turkey
bKilis 7 Aralık University, Business and Administration Department, Kilis, Turkey
cGaziantep University, Industrial Engineering Department, Gaziantep, Turkey
article info
Keywords:
Artificial Neural Network
Genetic Algorithm
Harmony Search Algorithm
Stock market price
abstract
Stock market price is one of the most important indicators of a country’s economic growth. That’s why de-
termining the exact movements of stock market price is considerably regarded. However, complex and un-
certain behaviors of stock market make exact determination impossible and hence strong forecasting models
are deeply desirable for investors’ financial decision making process. This study aims at evaluating the ef-
fectiveness of using technical indicators, such as simple moving average of close price, momentum close
price, etc. in Turkish stock market. To capture the relationship between the technical indicators and the stock
market for the period under investigation, hybrid Artificial Neural Network (ANN) models, which consist in
exploiting capabilities of Harmony Search (HS) and Genetic Algorithm (GA), are used for selecting the most
relevant technical indicators. In addition, this study simultaneously searches the most appropriate number
of hidden neurons in hidden layer and in this respect; proposed models mitigate well-known problem of
overfitting/underfitting of ANN. The comparison for each proposed model is done in four viewpoints: loss
functions, return from investment analysis, buy and hold analysis, and graphical analysis. According to the
statistical and financial performance of these models, HS based ANN model is found as a dominant model for
stock market forecasting.
© 2015 Elsevier Ltd. All rights reserved.
1. Introduction
A stock market is a public market to trade the company’s stocks
and derivative at an approved stock price (Preethi & Santhi, 2012).
Stock market provides opportunities for brokers and companies
to make investments on neutral ground and is one of the primary
indicators of a economic condition of the country (Perwej & Perwej,
2012). However, stock market is characterized by nonlinearities,
discontinuities, and high-frequency multi-polynomial components
because it is interacted with many factors such as political events,
general economic conditions, and traders’ expectations (Hadavandi,
Shavandi, & Ghanbari, 2010). Also, the fast data processing of these
events with the help of improved technology and communication
systems has caused the stock prices to fluctuate very fast. Thus many
banks, financial institutions, large scale investors and stock brokers
have to buy and sell stocks within the shortest possible time and time
span of even a few hours between buying and selling is not unusual
Corresponding author. Tel.: +90322 455 0000–2226.
E-mail addresses: mgocken@adanabtu.edu.tr (M. Göçken), mozcalici@kilis.edu.tr
(M. Özçalıcı), aboru@adanabtu.edu.tr (A. Boru), dosdogru@gantep.edu.tr
(A.T. Dosdo˘
gru).
(Bonde & Khaled, 2012). Robust and agile stock market is also highly
desirable in the field of finance, engineering and mathematics due
to high return possibility. It is generally seen as a peak investment
outlet. For these purposes, many researchers have been investigated
the predictability of the stock market by using of fundamental anal-
ysis, technical analysis, time series prediction, and machine learning
methods (Prasanna & Ezhilmaran, 2013). Besides, most of the com-
panies have created new methods for evaluating financial data and
investment decisions (Sureshkumar & Elango, 2012). Among them,
ANN approach has been thought as the best forecasting method
with a high level of validity in the fields of stock market forecasting.
However, some critical points of ANN structure should be carefully
analyzed. The definition what constitutes an optimal set of ANN
input variables can be considered one of the main problems in ANN
structure because the choice of input variables directly affects the
forecasting accuracy. Secondly, number of neurons (or units, nodes)
in hidden layer is also so important for ANN. It is an adjustable part
in ANN but unfortunately, there is no unique method for fixing the
optimum number of neurons in hidden layer for a particular problem.
Therefore, researchers prefer generally to use trial and error method
for this purpose. In this paper, we proposed hybrid methodology for
determining input variable and the number of neurons in hidden
http://dx.doi.org/10.1016/j.eswa.2015.09.029
0957-4174/© 2015 Elsevier Ltd. All rights reserved.
M. Göçken et al. /Expert Systems With Applications 44 (2016) 320–331 321
layer. GA and HS are used as a tool for improving ANN’s forecasting
performance. In literature, GA is often used with ANN for the purpose
of training the network, feature subset selection, and architecture
optimization. However, HS is generally not used with ANN for these
purposes. Therefore, our study is created alternative solution meth-
ods for stock market forecasting with better solutions. The contribu-
tion is structured as follows. Section 2 describestherelatedworks.
Then, we start describing solution methodology in Section 3.Section
4deals with results and discussions. Finally, Section 5 is devoted to
conclusions.
2. Literature
The importance of Turkish stock market has increased substan-
tially with the establishment of the Istanbul Stock Exchange in 1986.
Since its establishment in 1986, the ISE has followed a fast pace
growth in terms of trading volume, market capitalization, number
of listed corporations and foreign investment (Adaoglu, 2000). Also,
ISE characterized with high volatility in the market and such volatil-
ity attracts many local and foreign investors as it provides high re-
turn possibility (Cinko & Avci, 2009). Hence, forecasting stock market
movement has been the objective of the vast research papers apply-
ing different techniques. Among them, ANN is featured as being data
driven and, hence, does not require assumptions concerning data.
With such a feature ANN is a suitable technique in handling nonlin-
ear, highly complex and dynamic data of stock markets (Karymshakov
& Abdykaparov, 2012). In the literature, ANN is clearly explained
by Egeli, Ozturan, and Badur (2003). The authors used six different
ANNs which includes multi-layer perceptron (MLP) and generalized
feed forward to predict ISE market index value. Authors used previ-
ous day’s index value, previous day’s TL/USD exchange rate, previous
day’s overnight interest rate and 5 dummy variables each represent-
ing the working days of the week as inputs. The results showed that
for each ANN model, the highest accuracies were obtained with 1 hid-
den layer and also ANN models give more accurate results than the
ones based on moving averages. Guresen, Kayakutlu, and Daim (2011)
compared ANN models including MLP, dynamic ANN, and the hybrid
neural networks. It is observed that classical ANN model MLP gives
more reliable results than the other models used in this comparison.
Kara, Boyacioglu, and Baykan (2011) revealed that ANN works bet-
ter than Support Vector Machine in predicting the direction of stock
price movement in the ISE. In the study, parameters of ANN models
such as number of neurons in the hidden layer were determined em-
pirically. Also, ten technical indicators were selected as feature sub-
sets by the review of domain experts and prior researches. ¸Senol and
Özturan (2008) statistically demonstrated that ANN outperforms Lo-
gistic Regression methodology. In the study, ANN was used to pre-
dict the stock price behavior in terms of its direction. The best results
were obtained for ANN model with three inputs, 11hidden neurons
in the single hidden layer and one output with three indicators, rel-
ative strength index of 14 days, stochastic indicator for 14 days, and
stochastic moving average. Yildiz, Yalama, and Coskun (2008) utilized
ANN for forecasting the direction of the ISE National-100 using the
highest and lowest prices paid during the day, the closing price, the
exchange rate (as the US dollar), and response rates as an input vari-
ables. The results of the previous studies show that accuracy of stock
market prediction is generally between 60% and 76%, and hence more
robust ANN model is needed to increase prediction accuracy in Turk-
ish stock market.
Having highly functional stock markets and exchanges is incred-
ibly valuable all over the world is a well-known fact. Therefore, so
many types of ANN models are developed to search out more efficient
forecasting model. Chiu and Chuang (2003) showed that ANN has
ability for predicting tendency of Taiwan stock market. Five different
ANN models were developed to decide the number of input neurons
and hidden neuron. Also, the classification technique and clustering
method were used under framework of ANN with quantitative and
qualitative factors. Similarly, Aldin, Dehnavi, and Entezari (2012) used
ANN for stock price index forecasting on the Taiwan Stock Exchange.
Closing price, the high and low price index were converted into tech-
nical indicators for predicting the position of stock price movements.
In the study, neuron numbers in the hidden layer was determined
empirically. Dastgir and Enghiad (2012) evaluated Iran Stock Market
by focusing on forecasting Tehran Stock Exchange Price Index which
is the most significant index of Iran Stock Market. In the study, two
hidden layers were used with many combinations of architecture.
The number of neurons in each hidden layer was changed from one
to sixteen. Results of the study revealed that ANN model with three
hidden neurons on the first hidden layer and four hidden neurons
on the second achieved the best performance in Iran Stock Market.
Ruxanda and Badea (2014) presented different configured ANNs and
compared them in terms of forecasting errors while making predic-
tions on Bucharest Stock Market Index. Input variables were set based
on a stepwise forward regression. Adebiyi, Adewumi, and Ayo (2014)
found that 10 inputs obtained from the New York Stock Exchange
including open price, low price, high price, close price, and volume
traded, 17 hidden neurons, and one output neuron give more accu-
rate results in ANN model. Laboissiere, Fernandes, and Lage (2015)
used ANN to predict the maximum and minimum day stock prices
of Brazilian power distribution companies. In the study, correlation
analysis was used to select input variable and different ANN archi-
tectures were tested empirically. The best results were found with
one hidden layer and only five hidden neurons. Zahedi and Rounaghi
(2015) applied ANN and principal component analysis to predict
stock price on Tehran Stock Exchange. The results of the study show
that ANN model has superiority over its rivals. Also, principal com-
ponent analysis method can accurately predict stock price on Tehran
Stock Exchange using 20 accounting variables.
In this paper we review studies in the ANN literature which have
been used for stock market forecasting, results revealed that a dif-
ferent combination of attribute sets was experimented with differ-
ent ANN model parameter values and each study provides satisfying
result in existing condition but ANN architecture is very important
which directly affects system performance essentially. Hence, most
previous studies were focused on the improvement of the ANN archi-
tecture. However, there are few studies on the input variable selec-
tion from predetermined data set and there is no clear methodology
available for variable selection and determining number of hidden
neurons in hidden layer. Therefore, the basic idea that lies behind the
proposed models is not only selecting the most relevant input vari-
ables that are to be used by ANN models but also setting the number
of neurons in hidden layer by manipulating ANN structure via meta-
heuristics. Thus, proposed models based on GA and HS are applied to
improve forecasting accuracy and stability of ANN.
3. Solution methodology
3.1. Technical indicators
This section describes input variable selection methodology. For
each case, 45 technical indicators are considered as input variables.
Technical indicators are effective tools to characterize the real market
situation. Using technical indicators can be more informative than us-
ing pure prices (Nikfarjam, Emadzadeh, & Muthaiyah, 2010) and it is
very practical way for stock analysts and fund managers to analyze
stock market. On the other hand, this technique may not be a good
alternative solution for common investors because too many tech-
nical indicators are available to be considered as prediction factors
and the most commonly used technical indicators are ordinarily not
understandable. Therefore, selection of the useful technical indica-
tors accurately is the key issue to make a profit for those stock mar-
ket investors (Wei & Cheng, 2012). However, no method is successful
322 M. Göçken et al. /Expert Systems With Applications 44 (2016) 320–331
enough to consistently beat the market. Every stock index or stock
has unique characteristics. That means, say feature A might play an
important role in predicting future prices of stock X while feature
B might be regarded as redundant for that stock. For that reason it
seems that it is not possible to say that “feature A is a good predictor
for every stock”. Different features must be used for prediction at-
tempt in different time periods and/or different stocks. In our study,
technical indicators are applied as the input variables of ANN to fore-
cast the stock market index. GA and HS are integrated with ANN not
only for optimizing the architecture of ANN but also for determining
the indicators that has the most significant effect on the forecasting
performance. The underlying logic for using GA and HS for variable
selection is to evaluate the usefulness of indicators and eliminate ir-
relevant ones to simplify the proposed model. It should be noted that
there is no limit for the number of indicators to be considered by GA
and HS. In Table 1, all of the technical indicators considered in this
study together with the final indicator selection results of GA and HS
algorithms are illustrated. Note that, shaded variables are selected by
none of the optimization methods.
3.2. Artificial Neural Network
ANN is a computational network which attempt to simulate, in a
gross manner, the networks of nerve cell (neurons) of the biological
(human or animal) central nervous system (Graupe, 2007). The infor-
mation processing and physical structure of the brain is partially em-
ulated with a web of neural connections (Li, 1994) which has great ca-
pacity in modeling nonlinear systems. Also, ANN is known with good
generalization capabilities and is substantially robust against noisy
or missing data (Versace, Bhatt, Hinds, & Shiffer, 2004). On the other
hand, it is difficult to design ANN model for a particular forecasting
problem. Modeling issues should be considered carefully. Determin-
ing the appropriate architecture such as number of the input vari-
ables, hidden layers and hidden neurons in each layer can be consid-
ered as a critical factor (Vaisla & Bhatt, 2010). For example, number of
hidden layers and neurons in each hidden layer is proportional to the
ability of the network to approximate more complicated functions.
However, this does not infer that complicated structures of networks
will always perform better (Perwej & Perwej, 2012). If the network
has too many hidden neurons, it will follow the noise in the data due
to over parameterization leading to poor generalization for untrained
data (Subasi & Erçelebi, 2005). On the other hand, network with too
few hidden neurons would be incapable of differentiating between
complex patterns leading to only a linear estimate of the actual trend
(Kuru ¸s, Kılıç, & Uçan, 2013). To eliminate these hesitations, three dif-
ferent forecasting models are proposed in this study and their perfor-
mances are compared. The parameters of each proposed model are
given in Table 2.
From Table 2, it is apparently seen that as their name imply, the
parameters of the first two proposed models are set by using HS and
GA, respectively. Note that the third model does not employ any opti-
mization method. Thus, third model directly uses all considered fea-
tures for training the ANN. It should be noted that 10 neurons in the
hidden layer are selected arbitrarily for the third model. The general
architecture of proposed models is demonstrated in Fig. 1.
In Fig. 1, p is the input pattern, b1is the vector of bias weights
on the hidden neurons, and W1is the weight matrix between 0th
(i.e. input) layer and 1st (i.e. hidden) layer. a1is the vector con-
taining the outputs from the hidden neurons, and n1is the vec-
tor containing net-inputs going into the hidden neurons. a2is the
column-vector coming from the second output layer, and n2is the
column-vector containing the net inputs going into the output layer.
W2is the synaptic weight matrix between the 1st (i.e. hidden)
layer and the 2nd (i.e. output) layer and b2is the column-vector
containing the bias inputs of the output neurons. Each row of W2
matrix contains the synaptic weights for the corresponding output
Tabl e 1
Initial feature pool and final result of selection status.
HS GA
Technical indicators Is selected? Is selected?
(0: No) (0: No)
(1: Yes) (1: Yes)
1 Today’s closeprevious
close price
00
2 Previous close price 1 1
3 Previous highest price 1 1
4Previouslowestprice 1 1
5 Previous open price 0 0
6 5 day simple moving
average of close price
00
7 6 day simple moving
average of close price
01
8 10 day simple moving
average of close price
00
9 20 day simple moving
average of close price
00
10 5 day exponential moving
average of close price
00
11 6 day exponential moving
average of close price
10
12 10 day exponential moving
average of close price
11
13 20 day exponential moving
average of close price
11
14 5 day triangular moving
average of close price
11
15 6 day triangular moving
average of close price
01
16 10 day triangular moving
average of close price
11
17 20 day triangular moving
average of close price
11
18 Accumulation/distribution
oscillator
01
19 Close price moving average
convergence/divergence
01
20 9-period exponential
moving average of MACD
01
21 Acceleration opening price 0 1
22 Acceleration highest price 1 0
23 Acceleration lowest price 0 1
24 Acceleration close price 1 1
25 Momentum open price 1 1
26 Momentum highest price 1 1
27 Momentum lowest price 0 0
28 Momentum close price 0 1
29 Chaikin volatility 0 1
30 Fast stochastic %K00
31 Fast stochastic %D11
32 Slow stochastic %K00
33 Slow stochastic %D10
34 William’s %R00
35 Relative strength index 1 1
36 Bollinger middle band 1 1
37 Bollinger higher band 1 1
38 Bollinger lower band 1 0
39 Highest high 1 0
40 Lowest low 1 1
41 Median price 1 1
42 Price rate of change 1 0
43 Typical price 0 0
44 Weighted close 0 0
45 William’s
accumulation/distribution
00
neuron (Ahmed, Jafri, Ahmad, & Khan, 2007). Firstly, the neuron re-
ceives information from the environment and then this information
multiplied by the corresponding weights is added together and used
as a parameter within an activation (transfer) function (Haider &
Hanif, 2009). The transfer functions are used to prevent outputs from
reaching very large values that can ‘paralyze’ ANN structure (Duch
& Jankowski, 1999). For hidden layer, suitable transfer function is
M. Göçken et al. /Expert Systems With Applications 44 (2016) 320–331 323
Tabl e 2
Parameters of the models.
Parameters HS-ANN model GA-ANN model Regular ANN model
Size of the input layer Determined by HS =23 Determined by GA =26 45
Number of neurons in hidden layer Determined by HS =17 De te rm ined by GA =210
Transfer function in hidden layer Sigmoid tangent Sigmoid tangent Sigmoid tangent
Transfer function in output layer Pure-linear transfer function Pure-linear transfer function Pure-linear transfer function
Training function Levenberg–Marquardt training algorithm Levenberg–Marquardt training algorithm Levenberg–Marquardt training algorithm
Fig. 1. Architecture of the proposed neural network (Ahmed et al., 2007).
particularly needed to introduce non-linearity into the network be-
cause it gives the power to capture nonlinear relationship between
input and output (Ravichandran, Thirunavukarasu, Nallaswamy, &
Babu, 2005). In this part, tangent sigmoid transfer function is applied
in hidden layer. However, the use of sigmoid units at the outputs can
limit the range of possible outputs to the range attainable by the sig-
moid, and this would be undesirable in some cases (Bishop, 1995).
Hereby, a pure linear function is selected in output layer. The pure lin-
ear transfer function calculates the neuron’s output by simply return-
ing the value passed to it. After ANN model is constructed, training
of ANN is the next important step of the forecasting model. Training
of ANN is an iterative process like weights and bias of the network.
In this paper, proposed ANN-based forecasting models are trained
by Levenberg–Marquardt (LM) algorithm with optimum network pa-
rameters. LM is a trust region based method with hyper-spherical
trust region (Burney, Jilani, & Ardil, 2005) and is used as an intermedi-
ate optimization algorithm between the Gauss–Newton (GN) method
and gradient descent algorithm. Also, LM addresses the limitations of
each of those techniques (Kermani, Schiffman, & Nagle, 2005). When
the current solution is far from a local minimum, the algorithm be-
haves like a gradient descent method: slow, but guaranteed to con-
verge. When the current solution is close to a local minimum, it be-
comes GN method and exhibits fast convergence (Lourakis & Argyros,
2005). However, it is important to note that LM is very efficient when
training networks which have up to a few hundred weights (Hagan &
Menhaj, 1994).
3.3. GA-ANN forecasting model
When building an ANN, a number of parameters should be consid-
ered and unlimited ways are available to construct ANN. In the litera-
ture, particularly, input variable selection remains an important part
of ANN model development, due to the negative impact that poor se-
lection can have on the performance of ANNs during training and de-
ployment post-development (Ma ´
ndziuk & Jaruszewicz, 2011). In this
study, we used GA to overcome drawbacks of the input variable selec-
tion. GA is a general adaptive optimization search methodology based
on a direct analogy to Darwinian natural selection and genetics in bi-
ological systems (Huang & Wang, 2006). GA ensures the development
of new and better populations among different species during evo-
lution. Although most standard meta-heuristic algorithms used only
information from a single individual, GA uses information of a pop-
ulation of individuals (solutions) when they conduct their search for
better solutions (Pardalos,Pitsoulis,Mavridou,&Resende,1995). Ad-
ditionally, it is important to note that GA has proved its success in
search and optimization problems. Its ability to exploit the informa-
tion accumulated about an initially unknown search space in order
to bias subsequent searches into useful subspaces can be given as the
main reason for the success. This is the key feature, particularly in
large, complex, and poorly understood search spaces, where classical
search tools (enumerative, heuristic, etc.) are inappropriate, offering
a valid approach to problems requiring efficient and effective search
techniques (Martínez & Lozano, 2008).
Fig. 2 depicts a synthetic scheme of the GA based selection of in-
put variables. The advantage of the GA-ANN model lies in synergy
between the GA, which is used for the selection of the variables to be
used, and ANN, exploiting the selected variables. In the same manner,
GA is used to determine number of neurons in hidden layer because
inadequate neurons can restrict the relationship or too many neurons
can cause overtraining. Certainly, getting the correct balance between
numbers of neurons directly affect forecasting accuracy in models.
Basically, GA-ANN model is shown in Fig. 3 and can be summa-
rized as follows. It divides dataset as training and testing dataset. Fur-
thermore, training data set is also divided into the subsets to give the
ANN generalization ability. For this purpose we evaluated the can-
didate solutions on different subsets and obtained a mean of mean
squared error (MSE). The minimization of this error is performed by
GA. Calculating MSE continues until stopping criteria is satisfied. Also,
GA has several genetic operators that can be modified to improve the
performance of particular implementations, namely representation,
selection, crossover, and mutation. These procedures are given in the
next sections.
3.3.1. Chromosome representation
Chromosome representation is the first and the most important
operator obtained by encoding of a chromosome to represent a so-
lution. In literature, binary encoding is the most commonly used in
GA and gives many possible chromosomes even with small number
of alleles (Rajasekaran & Pai, 2003). Similarly, binary encoding is used
in this study and chromosome is considered to be composed of two
parts: (1) variable selection and (2) determination of the number of
hidden neurons (Fig. 4). If a variable is selected, gene is coded as 1,
otherwise 0. Similarly, if a node is selected, gene is coded as 1, oth-
erwise 0. Total length of the chromosome is determined as the sum
of the total number of variables considered and the total number of
neurons considered.
GA starts with a randomly generated initial population. Initial
population consists of a number of chromosomes that represent the
number of variables and the number of hidden neurons. After all fit-
ness values for the whole initial population are obtained, the chro-
mosomes evolve through successive iterations called generations. To
enhance diversity of the generation and to generate the population
of the next generation, GA operators such as selection, crossover and
mutation are activated.
3.3.2. Selection operator
Individuals, called parents, are selected based on a selection rule
to generate new, better solutions for next generations. In this study,
stochastic uniform selection which lays out a line is used. This
method chooses an individual according to its scaled fitness value.
324 M. Göçken et al. /Expert Systems With Applications 44 (2016) 320–331
Fig. 2. Flow-chart representing the GAs based variable selection system (Cateni, Colla, & Vannucci, 2011).
Fig. 3. GA-ANN iterative process for variable selection and determination of number of hidden layer neurons.
Fig. 4. Chromosome representation.
3.3.3. Crossover and mutation operators
In crossover operator, two chromosomes are randomly selected
and their chromosome strings are randomly cut to produce new
chromosomes. In this respect, a pair of parents is firstly randomly se-
lected from the mating pool. Secondly, a point, called crossover site,
along their common length is randomly selected, and the informa-
tion after the crossover site of the two parent strings are swapped,
thus creating two new children (Otman & Jaafar, 2011). An illustrative
crossover operator utilized in this study, is shown in Fig. 5.
Then, mutation operator is applied to provide a small amount
of random search. Without mutation, offspring chromosomes would
be limited to only the genes available within the initial population.
M. Göçken et al. /Expert Systems With Applications 44 (2016) 320–331 325
Fig. 5. Illustrative example of crossover operator.
Fig. 6. Illustrative example of mutation operator.
Tabl e 3
Parameters of the GA.
Elite count 2
Crossover fraction 0.8
# of generations 100
Population size 20
Mutation should be able to introduce new genetic material as well
as modify existing one (Fig. 6). With these new gene values, the GA
may be able to arrive at a better solution than was previously possible
(Kougias & Theodosiou, 2010).
After mutation operator, the candidate solutions obtained by GA
proceeds to the phase II (fitness function prediction). Best two indi-
viduals are saved for the next generations. This iterative process is
repeated over many generations. The run of GA terminates when the
termination criterion is satisfied. The best individual ever encoun-
tered during the run is typically designated as the result of the run.
The parameters of the GA are given in Table 3.
3.4. HS-ANN forecasting model
HS is based on the improvization process of musicians in a band.
In HS algorithm, multiple harmonics groups can be used in parallel.
Proper parallelism usually leads to better implantation with higher
efficiency (Geem, 2006). The good combination of parallelism with
elitism as well as a fine balance of intensification and diversification
is the key to the success of the HS algorithm, and in fact, to the suc-
cess of any meta-heuristic algorithms (Yang , 20 09). HS is simple in
concept, few in parameters, easy in implementation, imposes fewer
mathematical requirements. Therefore, HS has been successfully ap-
plied as an optimization method in many scientific and engineering
fields and was reported to be competitive alternative to many rivals
(Mahdavia, Fesanghary, & Damangir, 2007). In this paper, we pro-
posed HS-ANN model for determining the most relevant input vari-
ables and the number of neurons in hidden layer. The first step in HS-
ANN model is to divide dataset as training and testing dataset. Fur-
thermore, training data set is also divided into the subsets to give the
ANN generalization ability. For this purpose we evaluated the candi-
date solutions on 5 different subsets and obtained a mean of MSE. The
minimization of this error is performed by HS. Details of the proposed
HS-ANN model are shown in Fig. 7.
To apply HS, the problem should be formulated in the optimiza-
tion environment, having an objective function and constraints as
Eq. (1) (Mahdavia et al., 2007; Yadav, Kumar, Panda, & Chang, 2012):
Minimize (or Maximize)f(x)
Subject to xiXii=1,2,3,....,N(1)
where f(x) is the objective function with xas the solution vector com-
posed of decision variable xi,andXiis the set of a possible range of
values for each decision variable xi(LxiXiUxi), where Lxiand Uxi
are the lower and upper bounds for each decision variable, respec-
tively. In addition, the values of different parameters of the HS algo-
rithm also have to be specified. These parameters include harmony
memory size (HMS), harmony memory considering rate (HMCR),
pitch-adjusting rate (PAR).
3.4.1. Initialize the harmony memory (HM)
The initial HM consists of an HMS number of randomly gener-
ated solution vectors. Each component of the solution vector in HM is
initialized using the uniformly distributed random number between
the lower and upper bounds of the corresponding decision variable
[Lxi,Ux
i], where 1 iN.Theith component of the jth solution
326 M. Göçken et al. /Expert Systems With Applications 44 (2016) 320–331
Fig. 7. Flow-chart representing the ANNs based variable selection system.
vector is given by Eq. (2)
xj
i=Lxi+(UxiLxi).rand[0,1](2)
where j=1,2,3, …, HMS and rand[0, 1] is a uniformly distributed ran-
dom number between 0 and 1. Each row consists of a randomly gen-
erated solution vector, and the objective function value for the jth
solution vector is denoted by f(xj). The matrix formed is governed by
Eq. (3).
HM(j,1:N)=xj
HM(j,N+1)=fxj(3)
The HM with the size of HMS ×(N+1)can be represented by a
matrix, as
HM =
x1
1x1
2x1
3··· x1
Nf(x1)
.
.
.....
.
.
xHMS
1xHMS
2xHMS
3··· xHMS
Nf(xHMS)
In this study, each row of the matrix of HM coincides with a so-
lution. First value of each row gives the number of hidden layer neu-
rons. The rest gives the information of whether the considered vari-
able is selected or not. The last value of the row gives the objective
function of the related row (see Fig. 8).
M. Göçken et al. /Expert Systems With Applications 44 (2016) 320–331 327
Fig. 8. Representation of the HM matrix.
Tabl e 4
Parameters of the HS.
HMS 100
bw 0.2
HMCR 0.95
PAR 0. 3
Max iteration 10,000
3.4.2. Improvise a new harmony from the HM
Then, new harmony is improvised which is the essence of the
HS algorithm. In improvization, the HS generates a new harmony
vector, x=x
1,x
2,.....x
N, using the following rules: memory con-
sideration, pitch adjustment, and random selection. The original HS
algorithm consists of three operations for considering the computa-
tional intelligence or randomness as Eq. (4):
x
ixix1
i,x2
i,...,xHMS
i,(HMCR)
xiXi,(1HMCR)(4)
In this step, a random number is generated. If this value is less
than HMCR, value of 1 is chosen, else value of 0 is chosen. After the
memory consideration, each decision variable is evaluated to deter-
mine whether pitch adjustment is necessary or not. This evaluation
is carried out with PAR parameter which is the probability of pitch
adjusting and identified as Eq. (5):
x
i=xi±rand(0,1)×bw with probability PAR
xiwith probability (1PAR)(5)
where bw is the range of Xi, rand(0,1) is random number between 0–
1.Ifthevaluesare1inbothofPARandHMCR,thevalueischosen
for the new harmony. Other values are selected as 0. In this step, the
value of x
iis chosen randomly. The value of x
iis in the range of Xiand
it has probability of HMCR. Details are given in Fig. 9.
3.4.3. Generation of new HM
After selecting the new values, the objective function value is cal-
culated for new harmony vector. If this value is better than the worst
harmony vector in the harmony matrix, it is then included in the ma-
trix, while the worst one is taken out of the matrix. Then, harmony
memory matrix is sorted in descending order by the objective func-
tion value. These are repeated until the termination criterion which is
the pre-selected maximum number of cycles is satisfied. Parameters
oftheHSaregivenintheTable 4.
4. Results and discussions
The main purpose of this study is to propose new hybrid stock
price forecasting models to get more accurate and reliable forecast-
ing. In the first section, variables are determined for ANN model. In
Tabl e 5
Descriptive statistics of testing and training dataset.
Train dataset Test dataset
Mean 50082.53 74373.31
Standard deviation 14858.80 4494.85
Length (Days) 4000 160
regular ANN model, we used all the variables considered. However,
in GA-ANN and HS-ANN models; we reduce the input variable set to
an optimal subset. Among 45 relevant input variables, GA-ANN se-
lected 26 variables as the optimal input variable subset while HS-
ANN selected 23 input variables as the optimal variable subset. Sim-
ilarly, the optimal number of neurons in hidden layer is specified.
It should be noted that for both GA-ANN and HS-ANN forecasting
model, the number of hidden layer is considered to be 1. In regu-
lar model, 10 neurons in hidden layer are predetermined arbitrarily.
However, both GA-ANN and HS-ANN models characterized their own
number of neurons in the hidden layer. While GA is selected only 2,
HS has 17 neurons in hidden layer.
To construct ANN model; suitable training and testing samples
should also be selected. The first issue here is to split the data into two
separate sets, the training and testing data sets. Although there is no
general solution to this problem, several factors such as the problem
characteristic, the data type, and the size of the available data should
be considered in making this decision. In this study, price informa-
tion of BIST100 index between 08/06/2005 and 27/05/2013 (4000 ob-
servations) used as training dataset, and 28/05/2013 and 20/09/2013
(160 observations) used as testing the performance of forecasting
models. Details of training and testing data set are given in Table 5.
Nine different loss functions namely mean absolute error (MAE),
root mean square error (RMSE), mean absolute relative error (MARE),
mean squared relative error (MSRE), root mean squared relative error
(RMSRE), mean absolute percent error (MAPE), mean squared per-
centage error (MSPE), and root mean squared percentage error (RM-
SPE) are used to evaluate the performance of training and testing
data sets. Resulting values of these loss functions are summarized
in Table 6. Actual prices and predicted prices are compared for each
forecasting models.
According to the results, HS-ANN model outperformed other fore-
casting models in terms of all statistical loss functions. It should
be noted that regular ANN model produced the highest forecasting
errors. In general loss function values are smaller at training per-
formance. However this performance is not representative because,
training set is used in the training of the model. In order to assess
the performance of forecasting models, we have to test it with a new
dataset. This performance is called as the testing performance and is
a true indicator of the forecasting performance. Note that, all of these
indicators have the smaller-the-better characteristic.
328 M. Göçken et al. /Expert Systems With Applications 44 (2016) 320–331
Fig. 9. New harmony improvization concept for the proposed HS-ANN forecasting model.
Tabl e 6
Training and testing statistics of models.
ei=piaip: predicted
price, a: actual price
Training HS-ANN Training GA-ANN Training regular ANN Testing HS-ANN Testing GA-ANN Testing regular ANN
MAE 1
n
n
i=1
|ei|944.7658 978.9924 429.3262 2597.321 2950.251 2951.554
MSE 1
n
n
i=1
e2
i2475056 2916503 362116,2 11236305 12202954 14516650
RMSE 1
n
n
i=1
e2
i1573.231 1707.777 601.7609 3352.06 3493.273 3810.072
MARE 1
n
n
i=1
|ei
ai|0.019524 0.018384 0.008916 0.033814 0.038628 0.038191
MSRE 1
n
n
i=1
|ei
ai|20.000772 0.000681 0.000155 0.00181 0.001995 0.002256
RMSRE 1
n
n
i=1
|ei
ai|20.027787 0.026087 0.012447 0.042541 0.044671 0.047493
MAPE 100
n
n
i=1
|ei
ai|1.95244 1.838397 0.891558 3.381416 3.862837 3.819056
MSPE 100
n
n
i=1
|ei
ai|27.720985 6.805103 1.549378 18.09704 19.95454 22.55602
RMSPE 100
n
n
i=1
|ei
ai|22.778666 2.608659 1.24474 4.254061 4.46705 4.749318
Regular ANN produced the smallest errors on training dataset.
However in testing performance it produces the highest errors in al-
most all indicators except for MARE which means that regular ANN
memorizes the training set and lost the generalization ability. This
situation is also known as overfitting issue. When the performance of
HS-ANN and GA-ANN models are examined, it becomes clear that in
training dataset except for the first three indicators, HS-ANN model
produced higher errors in comparison to GA-ANN model. However,
in testing period HS-ANN model outperformed GA-ANN model which
means it produced less error. This also means that HS-ANN has better
generalization ability than GA-ANN model. Among the statistical loss
functions MAPE has the most human friendly characteristic and gives
the error in percent. Note that, HS-ANN model produced 3.38% error
which is an acceptable error rate.
Fig. 10 depicts actual and forecasted prices as a time series. Ev-
ery figure consists of two parts. In the upper side predicted and
M. Göçken et al. /Expert Systems With Applications 44 (2016) 320–331 329
020 40 60 80 100 120 140 160
6
7
8
9x 10
4
Actual Prices and HS ANN Forecasting Results
HS Predic ted
Actual Price
020 40 60 80 100 120 140 160
-15000
-10000
-5000
0
5000
Predicted - Actual Values
HS Predicted - Actual
020 40 60 80 100 120 140 160
6
7
8
9x 10
4
Actual Prices and GA ANN Forecasting Results
GA Predict ed
Actual Price
020 40 60 80 100 120 140 160
-10000
-5000
0
5000
Predicted - Actual Values
GA Predic ted - Act ual
020 40 60 80 100 120 140 160
6
7
8
9x 10
4
Actual Prices and Single ANN Forecasting Results
Single ANN Predicted
Actual Price
020 40 60 80 100 120 140 160
-10000
-5000
0
5000
Predicted - Actual Values
Single ANN Predicted - Actual
Fig. 10. Actual versus predicted prices as time series.
actual prices are illustrated while in the lower side predicted minus
actual values are shown. Difference values are useful to get a general
vision of how well the forecasting model is. It should be noted that
in a perfect prediction difference graph the values should lie around
zero. Since deviations from zero indicate a deviation from a good pre-
diction it could be said that HS-ANN prediction performance is better
than that of others. In GA and regular ANN models difference lines
are far away from zero line.
Statistical performance measurements do not have much meaning
for practical investors. Financial performance of forecasting model
must also be examined to evaluate forecasting model. We can simu-
late buying and selling behaviors of a typical investor. An investor will
buy stocks from the market if he/she expects an increase in prices.
Similarly if an investor anticipates a decrease in prices then he/she
will try to sell his/her financial assets in order to prevent a poten-
tial loss. We simulated the above mentioned simple trading logic us-
ing the prediction results (or predicted values) of the proposed mod-
els. A trading algorithm is developed to trade-on-paper along test-
ing period. This algorithm returned the paper profits obtained from
transactions. We neglected the trading costs and taxes to simplify the
calculations.
Similarly, financial performances of the proposed models are
compared with a passive trading strategy. In buy and hold strat-
egy, an investor buys stocks from the beginning of trading period
price and sells all of its assets from end of trading period closing
price. The return in percent from this transaction is calculated as
follow (Eq. (6)):
r=PtnPt
Ptn
(6)
Tabl e 7
A Comparison of the proposed models’ performances with a passive trading strategy.
HS-ANN model GA-ANN model Regular ANN model
Return from
investment
0.060406 0.011221 0.20626
Buy and hold 0.13405 0.13405 0.13405
where Ptnand Ptrepresents the first-day and last day closing price
of stock index in testing period.
The result of trading strategy yields a loss of 13.41%. HS-ANN
model yields a return of 6.04% profit while GA-ANN model returns
only 1.12% of profit during 160 trading sessions. By the way, regular
ANN yields a loss of 20.63% as can be seen from Table 7.
5. Conclusion
Over the years, researchers around the world have been studying
to forecast the stock market price as precisely as possible to reach
the best investment decisions. However, there is no consensus on the
effectiveness of forecasting models and hence, research on improv-
ing the effectiveness of forecasting models has been continued. This
paper has proposed a new hybrid model, based on a heuristic opti-
mization methodology (HS or GA) and ANN, to improve stock market
forecasting performance in terms of statistical and financial terms.
With development of the hybrid ANN models we show that struc-
turing ANN has become easy in implementation because our pro-
posed models have great capability in variable selection and deter-
mining the number of neurons in hidden layer. In order to select the
330 M. Göçken et al. /Expert Systems With Applications 44 (2016) 320–331
most relevant technical indicators, we firstly set predetermined 45
variables and at the end of the analysis 26 and 23 variable are speci-
fied as non-redundant by GA and HS models, respectively. That means
the complexity of the variable selection is reduced to almost its half.
In addition, determining the optimum number of neurons in hid-
den layer eliminates the overfitting or underfitting problems of ANN
models.
Based on the results, the average stock price forecasting perfor-
mance of the HS-ANN (MAPE =3.38) is significantly better than
that of GA-ANN (MAPE =3.86) model and the regular ANN model
(MAPE =3.81). It should be noted that MAPE values with pro-
posed models are about 10% lower than the ones reported in existing
studies. Furthermore, MAPE results with the proposed models look
promising in emerging markets. Also, trading performances are quite
impressive in proposed models. HS-ANN (%6.04) and GA-ANN (%1.12)
models yield higher returns in comparison with regular ANN model
(%20.63). Even operating in a bear market (buy and hold return is
%13.41), forecasting models accomplished to yield a positive return.
Although proposed hybrid models of predicting stock market
prices using the GA and HS give remarkable results, this study has
some limitations. First, number of hidden layer is fixed at 1. Al-
though training becomes excessively time-consuming with increas-
ing number of hidden layers, the performance of the model can
change with the number of hidden layer. The second limitation is pre-
determined transfer and training functions because combinations of
training function and transfer functions may affect quality of ANN
models.
HS-ANN and GA-ANN can be used successfully to forecast the
stock market price movement in different stock markets. The other
good direction for future research would be to consider other pa-
rameters which may affect the ANN architecture such as number of
hidden layer, type of transfer function. Variants of the HS such as
improved HS, global-best HS are possibly used to increase in the ac-
curacy of the forecasting. Similarly, the effects of various GA including
many different forms of selection, crossover, and mutation operators
can be examined as a part of the approach used in this study.
References
Adaoglu, C. (2000). Instability in the dividend policy of the Istanbul Stock Exchange
(ISE) corporations: evidence from an emerging market. Emerging Markets Review,
1(3), 252–270.
Adebiyi, A. A., Adewumi, A. O., & Ayo, C. K. (2014). Comparison of ARIMA and artificial
neural networks models for stock price prediction. Journal of Applied Mathematics,
1–7. doi:10.1155/2014/614342.
Ahmed, J., Jafri, M. N., Ahmad, J., & Khan, M. I. (2007). Design and implementation of a
neural network for real-time object tracking.World Academy of Science, Engineering
and Technology, International Journal of Computer, Electrical, Automation,Control and
Information Engineering, 1(6), 1816–1819.
Aldin, M. M., Dehnavi, H. D., & Entezari, S. (2012). Evaluating the employment of techni-
cal indicators in predicting stock price index variations using artificial neural net-
works (case study: Tehran Stock Exchange). International Journal of Business and
Management, 7(15), 25–34.
Bishop, C. M. (1995). Neural networks for pattern recognition. Oxford: Clarendon Press.
Bonde, G., & Khaled, R. (2012). Stock price prediction using genetic algorithms and evo-
lution strategies. In Proceedings of the 2012 international conference on genetic and
evolutionary methods (pp. 10–15).
Burney, S. M. A., Jilani, T. A., & Ardil, C. (2005). Levenberg–Marquardt algorithm for
Karachi Stock Exchange share rates forecasting. International Journal of Computa-
tional Intelligence, 1(3), 144–149.
Cateni, S., Colla, V., & Vannucci, M. (2011). A genetic algorithm-based approach for se-
lecting input variables and setting relevant network parameters of a SOM-based
classifier. International Journal of Simulation Systems, Science & Technology, 12(2),
30–37 UKSim 4th European modeling symposium on mathematical modeling and
computer simulation.
Chiu, D. Y., & Chuang, K. P. (2003). Applying artificial neural network and Chinese news
classification techniques to Taiwan Stock Market. Tamsui Oxford Journal of Mathe-
matical Sciences, 19(2), 201–215.
Cinko, M., & Avci, E. (2009). Examining the day of the week effect in Istanbul Stock
Exchange (ISE). International Business & Economics Research Journal, 8(11), 1–5.
Dastgir, M., & Enghiad, M. H. (2012). Short-term prediction of Tehran Stock Exchange
Price Index (TEPIX): using artificial neural network (ANN). Quarterly Journal of Se-
curities Exchange, 4(14), 237–261.
Duch, W., & Jankowski,N. (1999). Survey of neural transfer functions. Neural Computing
Surveys, 2(1), 163–212.
Egeli, B., Ozturan, M., & Badur, B. (2003). Stock market prediction using artificial neu-
ral networks. In Proceedings of the 3rd Hawaii international conference on business
(pp. 1–8). Hawaii, USA.
Geem, Z. W. (2006). Optimal cost design of water distribution networks using harmony
search. Engineering Optimization, 38(3), 259–280.
Graupe, D. (2007). Principles of artificial neural networks. Advanced series on circuits
and systems, World Scientific Publishing, Singapore City (Vol. 6) (2nd ed.).
Guresen, E., Kayakutlu, G., & Daim, T. U. (2011). Using artificial neural network models
in stock market index prediction. Expert Systems with Applications, 38(8), 10389–
10397.
Hadavandi, E., Shavandi, H., & Ghanbari, A. (2010). Integration of genetic fuzzy systems
and artificial neural networks for stock price forecasting.Knowledge-Based Systems,
23(8), 800–808.
Hagan, M. T., & Menhaj, M. B. (1994). Training feedforward networks with the Mar-
quardt algorithm. IEEE Transactions on Neural Networks, 5(6), 989–993.
Haider, A., & Hanif, M. N. (2009). Inflation forecasting in Pakistan using artificial neural
networks. Pakistan Economic and Social Review, 47(1), 123–138.
Huang, C. L., & Wang, C. J. (2006). A GA-based feature selection and parameters opti-
mization for support vector machines. Expert Systems with Applications, 31(2),231–
240.
Kara, Y., Boyacioglu, M. A., & Baykan, O. K. (2011). Predicting direction of stock price
index movement using artificial neural networks and support vector machines:
the sample of the Istanbul Stock Exchange. Expert Systems with Applications, 38(5),
5311–5319.
Karymshakov, K., & Abdykaparov, Y. (2012). Forecasting stock index movement with
artificial neural networks: the case of Istanbul Stock Exchange. Trakya University
Journal of Social Science, 14(2), 231–242.
Kermani, B. G., Schiffman, S. S., & Nagle, H. T. (2005). Performance of the Levenberg–
Marquardt neural networktraining method in electronic nose applications. Sensors
and Actuators B: Chemical, 110(1), 13–22.
Kougias, I., & Theodosiou, N. (2010). A new music-inspired harmony based optimiza-
tion algorithm. Theory and applications. In Proceedings of international conference
on protection and restoration of the environment X.Corfu.
Kuru ¸s, O. A., Kılıç, N., & Uçan, O. N. (2013). Hermitian transform approach in classifica-
tion of ECG signals. ˙
Istanbul Aydın Üniversitesi Dergisi, 2(7), 89–101.
Laboissiere, L. A., Fernandes, R. A. S., & Lage, G. G. (2015). maximum and minimum
stock price forecasting of Brazilian power distribution companies based on artifi-
cial neural networks. Applied Soft Computing, 35, 66–74.
Li, E. Y. (1994). Artificial neural networks and their business applications. Information
& Management, 27(5), 303–313.
Lourakis, M. I. A., & Argyros, A. A. (2005). Is Levenberg–Marquardt the most efficient
optimization algorithm for implementing bundle adjustment? In Proceedings of the
10th IEEE international conference on computer vision (pp. 1526–1531).
Mahdavia, M., Fesanghary, M., & Damangir, E. (2007). An improved harmony search al-
gorithm for solving optimization problems. Applied Mathematics and Computation,
188(2), 1567–1579.
Ma ´
ndziuk, J., & Jaruszewicz, M. (2011). Neuro-genetic system for stock index predic-
tion. Journal of Intelligent and Fuzzy Systems, 22(2), 93–123.
Martínez, C. G., & Lozano, M. (2008). Local search based on genetic algorithms,
advances in metaheuristics for hard optimization (pp. 199–221). Berlin Heidelberg:
Springer.
Nikfarjam, A., Emadzadeh, E., & Muthaiyah, S. (2010).Text mining approaches for stock
market prediction. In Proceedings of Computer and Automation Engineering (ICCAE):
4(pp. 256–260).
Otman, A., & Jaafar, A. (2011). A comparative study of adaptive crossover operators for
genetic algorithms to resolve the traveling s alesman problem. International Journal
of Computer Applications, 31(11), 49–57.
Pardalos, P. M., Pitsoulis, L., Mavridou, T., & Resende, M. G. C. (1995). Parallel search for
combinatorial optimization: genetic algorithms simulated annealing, tabu search and
grasp, parallel algorithms for irregularly structured problems (pp. 317– 331). Berlin
Heidelberg: Springer.
Perwej, Y., & Perwej,A. (2012). Prediction of the Bombay Stock Exchange (BSE) market
returns using artificial neural network and genetic algorithm. Journal of Intelligent
Learning Systems and Applications, 4(2), 108–119.
Prasanna, S., & Ezhilmaran, D. (2013). An analysis on stock marketprediction using data
mining techniques. International Journal of Computer Science & Engineering Technol-
ogy, 4(2), 49–51.
Preethi, G., & Santhi, B. (2012). Stock market forecasting techniques: a survey. Journal
of Theoretical & Applied Information Technology, 46(1), 24–30.
Rajasekaran, S., & Pai, G. A. V. (2003). Neural networks, fuzzy logic and genetic algorithm:
synthesis and applications (15th ed.) PHI Learning Pvt. Ltd.
Ravichandran, K. S., Thirunavukarasu, P., Nallaswamy, R., & Babu, R. (2005). Estimation
of return on investment in share market through ANN. Journal of Theoretical and
Applied Information Technology, 44–54.
Ruxanda, G., & Badea, L. M. (2014).Configuring artificial neural networks for stock mar-
ket predictions. Technological and Economic Development of Economy, 20(1), 116–
132.
¸Senol, D., & Özturan, M. (2008). Stock Price Direction prediction using artificial neural
network approach: the case of Turkey. Journal of Artificial Intelligence, 1(2), 70–77.
Subasi, A., & Erçelebi, E. (2005). Classification of EEG signals using neural network and
logistic regression. Computer Methods and Programs in Biomedicine, 78(2), 87–99.
Sureshkumar, K. K., & Elango, N. M. (2012). Performance analysis of stock price predic-
tion using artificial neural network. Global Journal of Computer Science and Technol-
ogy, 12(1), 18–26.
M. Göçken et al. /Expert Systems With Applications 44 (2016) 320–331 331
Vaisla, K. S., & Bhatt, A. K. (2010). An analysis of the performance of artificial neural
network technique for stock market forecasting. International Journal on Computer
Science and Engineering, 2(6), 2104–2109.
Versace, M., Bhatt, R., Hinds, O., & Shiffer, M. (2004). Predicting the exchange traded
fund DIA with a combination of genetic algorithms and neural networks. Expert
Systems with Applications, 27(3), 417–425.
Wei, L. Y., & Cheng, C. H. (2012). A hybrid recurrent neural networks model based on
synthesis features to forecast the Taiwan Stock Market. International Journal of In-
novative Computing Information and Control, 8(8), 5559–5571.
Yadav, P., Kumar, R., Panda, S. K., & Chang, C. S. (2012). An intelligent tuned harmony
search algorithm for optimization. Information Sciences, 196, 47–72.
Yang, X. S. (2009). Harmony search as a metaheuristic algorithm, in music-inspired
harmony search algorithm: theory and applications. In Z. W. Geem (Ed.), Studies in
computational intelligence (pp. 1–14). Berlin: Springer.
Yildiz, B., Yalama, A., & Coskun, M. (2008). Forecasting the Istanbul Stock Exchange
national 100 index using an artificial neural network. World Academy of Science,
Engineering and Technology, 22, 36–39.
Zahedi, J., & Rounaghi, M. M. (2015).Application of artificial neural network models and
principal component analysis method in predicting stock prices on Tehran Stock
Exchange. Physica A: Statistical Mechanics and Its Applications, 438, 178–187.
... So we chose to use 200 trees in the random forest of our simulations, since only modest gains in predictive performance are obtained past that point, and also not to create unnecessary computational burden. [46]. ...
Article
Full-text available
In this work, we propose a novel hybrid method for the estimation of regression models, which is based on a combination of LASSO-type methods and smooth transition (STR) random forests. Tree-based regression models are known for their flexibility and skills to learn even very nonlinear patterns. The STR-Tree model introduces smoothness into traditional splitting nodes, leading to a non-binary labeling, which can be interpreted as a group membership degree for each observation. Our approach involves two steps. First, we fit a penalized linear regression using LASSO-type methods. Then, we estimate an STR random forest on the residuals from the first step, using the original covariates. This dual-step process allows us to capture any significant linear relationships in the data generating process through a parametric approach, and then addresses nonlinearities with a flexible model. We conducted numerical studies with both simulated and real data to demonstrate our method’s effectiveness. Our findings indicate that our proposal offers superior predictive power, particularly in datasets with both linear and nonlinear characteristics, when compared to traditional benchmarks.
... Building on this foundation, our study leverages NN techniques to incorporate market-sentiment information into the stock-price prediction process. According to the literature, NN techniques fall into five main categories: (i) multiple-layer perceptron (MLP) (e.g., [15,37,94,102,113,116,121,138]), (ii) long short-term memory (LSTM) (e.g., [52,147]), (iii) convolutional NN (CNN) (e.g., [37,119]), (iv) recurrent NN (RNN) (e.g., [17,19]), and (v) heuristic (e.g., [89,96,138]). ...
Article
Full-text available
Price prediction tools play a significant role in small investors’ behavior. As such, this study aims to propose a method to more effectively predict stock prices in North America. Chiefly, the study addresses crucial questions related to the relevance of news and tweets in stock-price prediction and highlights the potential value of considering such parameters in algorithmic trading strategies—particularly during times of market panic. To this end, we develop innovative multi-layer perceptron (MLP) and long short-term memory (LSTM) neural networks to investigate the influence of Twitter count (TC), and news count (NC) variables on stock-price prediction under both normal and market-panic conditions. To capture the impact of these variables, we integrate technical variables with TC and NC and evaluate the prediction accuracy across different model types. We use Bloomberg Twitter count and news publication count variables in North American stock-price prediction and integrate them into MLP and LSTM neural networks to evaluate their impact during the market pandemic. The results showcase improved prediction accuracy, promising significant benefits for traders and investors. This strategic integration reflects a nuanced understanding of the market sentiment derived from public opinion on platforms like Twitter.
... In today's fiercely competitive environment, the vitality of market economies and share prices cannot be overstated, with the stock market serving as the pivotal axis (Göçken et al., 2016). Hence, acquiring a comprehensive understanding of financial and capital markets is imperative. ...
Article
Full-text available
The present study attempted to assess the impact of fundamental ratios on the share prices of selected telecommunication companies in India. India has dramatically expanded over the past ten years to become the second-biggest telecoms market worldwide, with 1.17 billion users. The Indian telecom industry has proliferated thanks in part to the government of India’s liberal and reformist policies and strong customer demand. It has become a lucrative investment sector for investors due to its recent and prospective growth. Data on 13 telecom firms indexed in the S&P BSE telecommunication index from 2013 to 2022 were taken from companies’ annual reports, the BSE website (Bombay Stock Exchange), and other secondary sources. Six firm-specific fundamental factors viz. Debt to Equity ratio (D/E), Current ratio (CR), Total Assets Turnover ratio (ATR), Earnings per share (EPS), Price to earnings ratio (P/E), Return on equity (ROE), and three country-specific fundamental factors viz. Gross Domestic Product, Inflation rate, and S&P BSE Sensex return were considered. Fixed effect panel regression through Generalized Least Square (GLS) model was performed to find inferences. Debt Equity ratio and Inflation rate were found to impact share price negatively. Conversely, the Total Assets Turnover ratio (ATR), Earnings per share (EPS), Price to Earnings ratio (P/E), and Return on Equity (ROE) positively impacted selected companies’ share prices. The study results will benefit individual & institutional investors in formulating their investment and portfolio diversification strategies for gaining a high effective rate of return on their investments.
... On the other hand, artificial neural networks (ANNs) are wellsuited for detecting relationships and nonlinear patterns in data that conventional linear models may overlook (Göçken et al. 2016;Haykin 1998). However, due to their black box nature, they are inappropriate for examining causal relationships and testing hypotheses (Chan and Chong 2012). ...
Article
Effective knowledge management (KM) and knowledge sharing (KS) are critical for the success of international construction projects (ICPs), but they are often hindered by knowledge-hiding (KH) behaviour. This study examines the complex and multifaceted factors that affect KH in ICPs, informed by Role Identity Theory (RIT) and Social Exchange Theory (SET). In particular, we investigate how ethnic minorities (EMs), contingent workers (CWs), and interpersonal justice (IJ) influence KH in these projects. Using structural equation modelling (SEM) and artificial neural networks (ANNs), we reveal the significant effects of EMs and CWs on KH and the partial mediation of IJ between EMs and KH. We also find that control variables such as gender and age have significant impacts on KH. The findings underscore the pivotal role of inclusive practices in mitigating KH and enhancing KM in ICPs. This research provides actionable insights for creating a transparent and inclusive knowledge-sharing culture in multicultural project teams, which contributes to improved project outcomes and organisational success. This study also paves the way for future research on power asymmetries, contextual factors, and the influence of organisational cultures, leadership styles, and team dynamics on KM in ICPs. Practical Applications: To address the challenges of knowledge hiding in international construction projects, this study proposes a comprehensive strategy for effective knowledge management in international construction projects, based on established research and best practices. The strategy is designed to fit the unique dynamics of international construction projects and includes the following components: Implementing a knowledge-sharing framework that values the contributions of each team member, regardless of their demographic backgrounds; Supporting leadership with a focus on inclusivity and equity, and a reward system acknowledging knowledge-sharing efforts irrespective of demographic backgrounds; Investing in construction-specific knowledge-sharing technologies that are accessible to all team members; Fostering inclusive communities of practice and incorporating performance metrics for inclusive knowledge sharing; Encouraging open communication, developing knowledge transfer plans, and ensuring continuous improvement in overall knowledge management. This multi-faceted approach aims to create a transparent and inclusive culture of knowledge sharing in international construction projects, which contributes to improved project outcomes and organisational success.
... AI has the potential to predict the future demand and supply of the market based on massive past data. Although its prediction and calculation may not be 100% accurate depending on the model section and data inputs (Göçken et al., 2016), it could make the distribution easier and more scientific. This is not to suggest that algorithms could make decisions alone or propose a pre-planned economy. ...
Preprint
Full-text available
Abstract This dissertation aims to explore the feasibility of a socialist economy that can better emancipate productivity and reduce inequality, considering the limitations of capitalism exposed by the COVID-19 pandemic and recent developments in algorithm development. Specifically, it investigates whether and how a socialist model that combines publicly owned means of production with market competition can be designed and implemented in light of recent developments in internet platforms, big data algorithms, and AI systems. The dissertation begins by examining the major feasibility issues of socialism, namely price signal and calculative knowledge of the market, as well as the competition and innovation incentive issue discussed in the well-known 1930 debates. It reviews past socialism approaches, including market socialism and techno-socialism, and considers the revolutionary momentum of the internet and recent AI algorithm development. It addresses how the capitalist system may be inadequate to such technological developments and stresses the need for public ownership in the current world. The dissertation proposes a new socialism approach that aimed to utilize a well-designed and tech- supported algorithm to determine the share of profit that should be allotted to workers and retired workers as lifelong income, as well as the proportion that should be reinvested and innovated to enhance productivity. This approach would preserve the link between individual labor input and personal income. While discussing the positive and important role that AI algorithms could play in such future governance and economic structures, it also acknowledges the issues that still exist, including data sources and concentration of public power. To conclude, while with limitations to be further studied, the embrace of algorithm-based governance and market competition may take the feasibility issue of socialism to a new page and support a renewed social system that could better adapt to current world.
Article
Full-text available
This paper presents a novel approach to constructing a load surrogate model. The surrogate estimates the damage equivalent loads (DELs) of a wind turbine of a given type regardless of its position within a wind farm, and under various farm control actions. The model relies solely on local inflow quantities (sector-averaged wind speeds and turbulence intensities) and local control parameters (rotor speed, pitch angle, and yaw misalignment). Despite its highly simplified representation of the complex behavior of the turbulent wind field, wake effects, and controller dynamics, these quantities prove sufficient to characterize DELs. The paper demonstrates the training of this load model within a simulation environment. Validation results using a different wind farm configuration indicate that the surrogate can accurately predict fatigue loads for both unwaked and waked turbines, encompassing scenarios of wake steering and induction control.
Article
The purpose of this research is to build an estimated model for the ecological footprint (EF) in the G20 countries in the period of 1999–2018, the last two decades. These countries have faced extensive EF changes and developments. This aim, indices (The Global Footprint Network and World Bank) has been used. Artificial neural network (ANN) models have been used for data analysis and index fitting. The findings of this research showed that the EF has been increasing in China with an average of (16.37), France (14.58), Brazil (9.37), and the United States (3.66). Root mean square error value in the first model (0.140), in the second model (0.0275), in the third model (0.0275) in the fourth model, is equal to (0.0608) and in the fifth model it is equal (0.11484). Therefore, the accuracy of EF prediction in neural network models 2 and 4 is 97.5% and 97.5%, respectively. Also, we believe that in better management of the EF, the use the ANN can be efficient and effective in forecasting the G20 countries. This study can provide two important contributions to the managers of this country in examining the environmental degradation of the G20 countries. First, the use of ANN is one of the methods of artificial intelligence that has a very high and acceptable accuracy in selecting research variables. Second, we used ANN models to solve the problem of prediction accuracy and selection of effective variables.
Article
Predicting agricultural commodity prices accurately is of utmost importance due to various factors such as perishability, seasonality, production uncertainty etc. Moreover, the substantial volatility that may be exhibited in time series further adds to the complexity and constitutes a significant challenge. In this paper, a Hidden Markov (HM) guided Deep Learning (DL) models has been developed on nonlinear and nonstationary price data of agricultural commodities for forecasting by considering technical indicators viz., Moving Average (MA), Bollinger Bands (BB), Moving Average Convergence Divergence (MACD), Exponential MA (EMA) and Fast Fourier Transformation (FFT). HM Models (HMMs) can effectively handle the sequential dependencies and hidden states, while DL approach can learn complex patterns and relationships within the price series and thus the drawback of lack of generalization capability in the DL model has been overcome by HMM. In this study, the Potato price data of the Champadanga district of West Bengal, India has been utilized to assess the performance of the proposed technique. HMM has been combined with six baseline DL models viz., Recurrent Neural Networks (RNN), Convolutional Neural Networks (CNN), Long Short-Term Memory (LSTM), Gated Recurrent Units (GRU), Bidirectional LSTM (BiLSTM) and Bidirectional GRU (BiGRU) for forecast modeling. Performance evaluation metrics viz., Root Mean Squared Error (RMSE), Mean Absolute Percentage Error (MAPE), Mean Absolute Error (MAE) and the insightful Diebold–Mariano (DM) test revealed that Hidden Markov hybridized with DL models surpassed baseline DL models in forecasting accuracy for 1-week, 4-week, 8-week and 12-week ahead DL predictions. The proposed approach holds significant promise for enhancing the precision of agricultural commodity price forecasting with far-reaching implications for various stakeholders such as farmers and planners.
Article
Full-text available
Dividend policy behaviour of corporations operating in emerging markets is significantly different from the widely accepted dividend policy behaviour of corporations operating in developed markets. This study provides evidence from the Istanbul Stock Exchange (ISE), an emerging European stock market, and analyses empirically whether the ISE corporations follow stable cash dividend policies in a regulatory environment that imposed mandatory dividend policies. Unlike the empirical results supporting the stable dividend policy behaviour of corporations operating in developed markets, the empirical results show that the ISE corporations follow unstable cash dividend policies and the main factor that determines the amount of cash dividends is the earnings of the corporation in that year.
Article
Full-text available
Genetic algorithm includes some parameters that should be adjusting so that the algorithm can provide positive results. Crossover operators play very important role by constructing competitive Genetic Algorithms (GAs). In this paper, the basic conceptual features and specific characteristics of various crossover operators in the context of the Traveling Salesman Problem (TSP) are discussed. The results of experimental comparison of more than six different crossover operators for the TSP are presented. The experiment results show that OX operator enables to achieve a better solutions than other operators tested.
Article
Full-text available
The choice of transfer functions may strongly influence complexity and performance of neural networks. Although sigmoidal transfer functions are the most common there is no apriorireason why models based on such functions should always provide optimal decision borders. A large number of alternative transfer functions has been described in the literature. A taxonomy of activation and output functions is proposed, and advantages of various non-local and local neural transfer functions are discussed. Several less-known types of transfer functions and new combinations of activation/output functions are described. Universal transfer functions, parametrized to change from localized to delocalized type, are of greatest interest. Other types of neural transfer functions discussed here include functions with activations based on nonEuclidean distance measures, bicentral functions, formed from products or linear combinations of pairs of sigmoids, and extensions of such functions making rotations...