Yeonguk Yu* and Yoon-Joong Kim*Two-Dimensional Attention-Based LSTM Model for Stock Index PredictionAbstract: This paper presents a two-dimensional attention-based long short-memory (2D-ALSTM) model for stock index prediction, incorporating input attention and temporal attention mechanisms for weighting of important stocks and important time steps, respectively. The proposed model is designed to overcome the long-term dependency, stock selection, and stock volatility delay problems that negatively affect existing models. The 2D-ALSTM model is validated in a comparative experiment involving the two attention-based models multi-input LSTM (MI-LSTM) and dual-stage attention-based recurrent neural network (DARNN), with real stock data being used for training and evaluation. The model achieves superior performance compared to MI-LSTM and DARNN for stock index prediction on a KOSPI100 dataset. Keywords: Attention Mechanism , LSTM , Stock Index Prediction , Two-Dimensional Attention 1. IntroductionPredictions of stock prices are a challenge due to the fact that stock prices are non-stationary and nonlinear at a certain point in time. In recent years, several studies have applied deep learning approaches to solve the challenge with promising results. The main deep learning models applied to predict stock volatility are convolutional neural network (CNN), recurrent neural network (RNN), long short-term memory (LSTM), and attention architecture, which are commonly utilized in the areas of image processing and neural machine translation. Cao and Wang [1] exemplified the use of CNN in this area by proposing a stock price forecasting model based on a modified CNN. Further, LSTM has become the state-of-the-art model for sequential data application. Greff et al. [2] investigated and analyzed its various variants and concluded that there were no better variants than the standard LSTM. Saad et al. [3] exploited time delay, recurrent, and probabilistic neural networks for predictability analysis, and concluded that all the networks are feasible. As regards attempts to add financial news or refine data features, Berradi and Lazaar [4] used featuredimensionality reduction to enhance the accuracy of the RNN model, which resulted in good prediction for stock prices. Minh et al. [5] used both financial news and a sentiment dictionary to predict the directions of stock prices based on the modified gated RNN. Several researchers have also suggested new ideas, and modified the structure of LSTM to accommodate those ideas [5,6]. In one such case, Baek and Kim [6] devised an overfitting prevention LSTM module and a prediction LSTM module. In the above stock price prediction model based on LSTM, the stock price sequence is transformed into the temporal context in a single fixed-length vector, which is fed to the network to produce the predicted stock price of the next day. The performance of the model varies depending on how much information the vector can capture from the source sequence. However, the data on the front-end of the sequence is diminished in the vector representation as the sequence enlarges, which results in decreasing performance—which is commonly referred to as the “long-term dependency problem (LTDP)”. To overcome this problem, researchers have proposed an attention mechanism for neural machine translation (NMT) [7], which extends to various fields including stock-related applications [8-10]. Liu and Wang [8] suggested an attention-based method using news and numerical data in stock market prediction. Li et al. [9] identified the stock selection problem (SSP), which is the issue whereby selection of stock prices weakly correlated with the target stock price, can degrade the prediction performance. They then proposed the attention-based multi-input LSTM (MI-LSTM), which can select important stocks from positively and negatively correlated stocks and demonstrated that the MI-LSTM models outperform other stock prediction models using LSTM. Furthermore, Qin et al. [10] proposed the dualstage attention-based RNN (DARNN), drawing inspiration from the encoder-decoder structure used in machine translation. The DARNN model predicts the stock index of the next day using past stock indexes and prices as input. This model consists of an attention-based encoder and decoder. The attention-based encoder is composed of an LSTM and input attention mechanism that applies attention weights to all stock according to their importance. The encoder output is taken as the attention-based decoder input. The attention-based decoder is composed of an LSTM and temporal attention mechanism that applies attention weights across all time steps for selection of relevant time steps. In other words, Qin et al. [10] proposed the use of an attention-based encoder and decoder to mitigate the problems of stock selection and long-term dependency by allowing the model to focus on informative stocks and times steps. They also demonstrated that their DARNN model outperforms traditional RNN models for stock index prediction. The occurrence of a stock-market-related event can affect stock volatility and is reflected in stock prices, and the extent and timing of the changes may vary depending on the stock. We refer to this characteristic of stock-dependent event-driven volatility as the stock volatility delay problem (SVDP). To solve this problem, a stock price prediction model must learn the corresponding temporal change patterns and employ them for prediction. However, DARNN [10] cannot solve the stock volatility delay problem, as the temporal attention mechanism uses the output of the input attention mechanism that applies attention weights for selection of relevant stock. Specifically, the output of the input attention mechanism is modified according to the stock importance; therefore, in the output, the temporal information that reflects the stock volatility pattern is distorted. To prevent this distortion, during calculation of the temporal-attention attention weights, it is necessary to use original stock information for which the stock change pattern has not been distorted by weight-based stock selection. Therefore, in this paper, we propose the two-dimensional attention-based LSTM (2D-ALSTM), which allows the same source sequence to be used for calculating the alignment scores of both the input attention and temporal attention mechanisms. Thus, this model can overcome the stock selection, longterm dependency, and stock volatility delay problems, which are negatively affected in attention-based models such as MI-LSTM and DARNN. The remainder of this paper is organized as follows. In Section 2, we introduce traditional LSTM, the attention mechanism, and the problem definition. In Section 3, we present and describe our proposed model. In Section 4, we report the results of a 2D-ALSTM model evaluation experiment, which are compared with results for MI-LSTM and DARNN. Finally, Section 5 gives the conclusion. 2. PreliminariesIn this section, we introduce the LSTM and attention mechanism, which form the foundation of all models considered in this study, i.e., MI-LSTM, DARNN, and 2D-ALSTM. Then, we define the prediction problem we consider for model evaluation. 2.1 Long Short-Term MemoryLSTM is one of the most commonly employed RNNs for time series prediction. The LSTM cell state [TeX:] $$\mathrm{h}_{t}$$ and hidden state [TeX:] $$C_{t}$$ are controlled by three gates: the output, input, and forget gates. For a given input time sequence [TeX:] $$X=\left(x_{1}, x_{2}, \cdots, x_{T}\right), \text { where } x_{t} \in \mathbb{R}^{m}$$ represents the input at time step t and m is the input size, the LSTM updates states are as follows:
Here, [TeX:] $$h_{t} \text { and } h_{t-1} \in \mathbb{R}^{q}$$ are the cell states at time steps t and t1, respectively; q is the number of LSTM hidden units; and [TeX:] $$\left[h_{t-1} ; x_{t}\right] \in \mathbb{R}^{m+q}$$ is a concatenation of [TeX:] $$h_{t-1} \text { and } x_{t}.$$ Furthermore, [TeX:] $$W_{f}, W_{i}, W_{j}, \text { and } W_{o} \in \mathbb{R}^{m \times(m+q)}$$ are weight parameters to be learned, while [TeX:] $$b_{f}, b_{i}, b_{j}, \text { and } b_{o} \in \mathbb{R}^{q}$$ are bias parameters to be learned. Finally, represents the sigmoid function. For convenience, we use a simple function notation [TeX:] $$f_{l s t m}$$ to represent the LSTM calculation described in Eqs. (1)–(6):
2.2 Attention MechanismThe attention mechanism was proposed to solve the long-term dependency problem encountered for the sequence-to-sequence (S2S) model used in machine translation. The S2S model is composed of two LSTMs called the encoder and decoder. The encoder converts input sequences of various lengths into information expressed in a single context vector using the LSTM; the decoder then uses this context vector to produce output sequences. However, when the input sequence length becomes excessive, a longterm dependency problem arises for the S2S model, as the encoder does not properly transmit the frontend information of the input sequence to the decoder. As noted above, use of an attention mechanism has been proposed to overcome the long-term dependency problem [7]. The attention mechanism causes the decoder to focus on the important encoder LSTM output when calculating its own output. Thus, the decoder input [TeX:] $$C_{t}$$ at t is calculated based on the previous decoder cell state [TeX:] $$S_{t-1}$$ and the encoder cell states [TeX:] $$\mathrm{H}=\left(h_{1}, h_{2}, \cdots, h_{T}\right)$$ as follows:
(10)[TeX:] $$a_{t, i}=\frac{\exp \left(\operatorname{score}\left(s_{t-1}, h_{i}\right)\right)}{\sum_{j=1}^{T} \exp \left(\operatorname{score}\left(s_{t-1}, h_{j}\right)\right)}$$Thus, as shown in Eq. (9), [TeX:] $$c_{t}$$ is the weighted sum of [TeX:] $$h_{i}.$$ Note that Eq. (10) shows calculation of the attention weight [TeX:] $$a_{t, i}$$ using the softmax function and score function. With the help of the attention mechanism, the dependencies between source and target sequences are not restricted by the in-between distance. Consequently, the attention mechanism was soon extended into various fields, including that of stock price prediction [8-10]. 2.3 Notation and Problem StatementIn this paper, we propose a model for time series prediction and evaluate this model using a stock index prediction task. For stock index prediction, we take a stock index sequence and stock price sequence as the model input. We refer to the stock index sequence as [TeX:] $$y=\left(y_{1}, y_{2}, \cdots, y_{T}\right) \in \mathbb{R}^{T},$$ where T is the time window size. In addition, we refer to the stock price sequence as [TeX:] $$X=\left(\mathbf{x}_{\mathbf{1}}, \mathbf{x}_{\mathbf{2}}, \cdots, \mathbf{x}_{T}\right) \in \mathbb{R}^{n \times T},$$ where n is the number of stocks. Hence, [TeX:] $$x_{t}=\left(x_{t}^{1}, x_{t}^{2}, \cdots, x_{t}^{n}\right) \in \mathbb{R}^{n}$$ is the price sequence of all stock at time t, and [TeX:] $$x^{k}=\left(x_{1}^{k}, x_{2}^{k}, \cdots, x_{T}^{k}\right) \in \mathbb{R}^{T}$$ is the price sequence of the k-th stock. The goal of the model is to predict the stock index of the next day, for a given y and X. We use [TeX:] $$\bar{y}_{T+1}$$ to represent the predicted real index of the next day given by the model, where [TeX:] $$y_{T+1}$$ represents the actual value. Thus, our model can be defined as follows:
3. Proposed ModelTo mitigate the performance degradation caused by problems such as stock selection, long-term dependency, and stock volatility, we propose a novel 2D-ALSTM model that consists of a CNN layer, an input and temporal attention layer, and an LSTM RNN layer, as shown Fig. 1. We use a CNN to transform the source stock price sequences to the effective feature sequences, which are appropriate for determining the 2D-attention (input and temporal attention mechanisms) attention weights. Finally, the context vector is calculated by the 2D-attention component, and we concatenate the context vector and index for use as the LSTM input. This process is described in more detail in the following subsections.3.1 CNN of 2D-ALSTMThe CNN produces the feature matrix D for a given stock price sequences X, as shown in Fig. 1. The CNN with filter size 5 is applied to the k stock price sequences through the time axis to produce the feature [TeX:] $$d_{t}^{k},$$ which forms the stock feature sequence [TeX:] $$d^{k}$$ and the temporal feature sequence [TeX:] $$d_{t},$$ such that we obtain D as follows for all stock:
(12)[TeX:] $$\boldsymbol{D}=\left(\boldsymbol{d}^{1}, \boldsymbol{d}^{2}, \ldots, \boldsymbol{d}^{\mathrm{n}}\right)$$
(13)[TeX:] $$\boldsymbol{d}^{k}=\left(d_{1}^{k}, d_{2}^{k}, \cdots, d_{T}^{k}\right) \in \mathbb{R}^{T}$$
(14)[TeX:] $$\boldsymbol{d}_{t}=\left(d_{t}^{1}, d_{t}^{2}, \cdots, d_{t}^{n}\right) \in \mathbb{R}^{n}$$
(15)[TeX:] $$d_{t}^{k}=\left(W_{d}\left[x_{t-2}^{k} ; x_{t-1}^{k} ; x_{t}^{k} ; x_{t+1}^{k} ; x_{t+2}^{k}\right]+b_{d}\right)$$where [TeX:] $$W_{d} \in \mathbb{R}^{1 \times 5}$$ is the weight parameter for training and [TeX:] $$b_{d} \in \mathbb{R}^{1}$$ is the bias. 3.2 Input Attention Mechanism of 2D-ALSTMThe input attention mechanism of the 2D-ALSTM is shown in the section labeled a) of Fig. 1. We use the stock feature sequence [TeX:] $$d^{k}$$ and the previous LSTM cell states [TeX:] $$h_{t-1} \text { and } h_{t-2}$$ as the alignment score function input. The score function applies linear regression to [TeX:] $$\boldsymbol{d}^{k}$$ and the concatenated state [TeX:] $$\left[h_{t-1} ; h_{t-2}\right]$$ and uses query vector [TeX:] $$v_{\mathrm{e}}$$ to calculate the alignment score [TeX:] $$e_{t}^{k}.$$ Following calculation of the [TeX:] $$e_{t}^{k}$$ values for all stock, the softmax function is applied to convert [TeX:] $$e_{t}^{k}$$ to attention weight [TeX:] $$a_{t}^{k}$$. We obtain [TeX:] $$\overline{X_{t}},$$ which is the input-attention output for all stock, by multiplying [TeX:] $$a_{t}^{k}$$ by [TeX:] $$x^{k}$$ as follows:
(16)[TeX:] $$e_{t}^{k}=v_{e}^{\top} \tanh \left(W_{e}\left[h_{t-1} ; h_{t-2}\right]+W_{i n d i c} d^{k}\right)$$
(17)[TeX:] $$a_{t}^{k}=\frac{\exp \left(e_{t}^{k}\right)}{\sum_{j=1}^{n} \exp \left(e_{t}^{j}\right)}, 1 \leq k \leq n$$
where [TeX:] $$v_{e} \in \mathbb{R}^{\mathrm{T}}, W_{e} \in \mathbb{R}^{\mathrm{T} \times 2 \mathrm{m}}, \text { and } W_{\text {indic}} \in \mathbb{R}^{\mathrm{T} \times \mathrm{T}}$$ are the weight parameters to be trained. We omit the bias for succinctness. The alignment score [TeX:] $$e_{t}^{k}$$ indicates the important score of the k-th stock feature sequence [TeX:] $$d^{k}$$ for the LSTM state [TeX:] $$h_{t}.$$ As the normalized score [TeX:] $$a_{t}^{k}$$ is a probability variable of the importance probability distribution over the stocks, the input-attention output [TeX:] $$\overline{\boldsymbol{x}}^{k}$$ represents the updated stock price of the k-th stock [TeX:] $$x^{k}$$ weighted by the importance score [TeX:] $$a_{t}^{k},$$ and thus constitutes data for solution of the stock selection problem. 3.3 Temporal Attention Mechanism of 2D-ALSTMWe use the temporal attention to obtain the context vector [TeX:] $$u_{t},$$ as shown in the section label ‘b’ in Fig. 1. We use the temporal feature sequence [TeX:] $$\boldsymbol{d}_{i}, \quad h_{t-1} \text { and } h_{t-2}$$ for linear regression. We perform a query with query vector [TeX:] $$v_{r}$$ to obtain the alignment score [TeX:] $$r_{t}^{i},$$ which indicates the important score of input [TeX:] $$\boldsymbol{d}_{i}$$ at time i for the LSTM state [TeX:] $$h_{t}.$$ Then, the softmax function is applied to obtain the attention weight [TeX:] $$\beta_{t}^{i},$$ which represents the probability variable of the importance probability distribution over the time. Finally, we obtain the context vector [TeX:] $$u_{t}$$ for the LSTM state [TeX:] $$h_{t}$$ by multiplying [TeX:] $$\beta_{t}^{i} \text { by } \bar{x}_{i}$$ of the input-attention output at time i. This procedure can be summarized as follows:
(19)[TeX:] $$r_{t}^{i}=v_{r}^{\mathrm{T}} \tanh \left(W_{e}\left[h_{t-1} ; h_{t-2}\right]+W_{t i m e} \boldsymbol{d}_{i}\right)$$
(20)[TeX:] $$\beta_{t}^{i}=\frac{\exp \left(r_{t}^{i}\right)}{\sum_{j=1}^{T} \exp \left(r_{t}^{j}\right)}$$
(21)[TeX:] $$\boldsymbol{u}_{\boldsymbol{t}}=\sum_{j=1}^{T} \beta_{t}^{j} \overline{\boldsymbol{x}}_{j}$$where [TeX:] $$v_{r} \in \mathbb{R}^{n}, W_{r} \in \mathbb{R}^{\mathrm{n} \times 2 \mathrm{m}}, \text { and } W_{t i m e} \in \mathbb{R}^{\mathrm{n} \times \mathrm{n}}$$ are the weight parameters for training. We omit the bias for succinctness. The role of the alignment score [TeX:] $$r_{t}^{i}$$ is to allow the model to compute the importance score of i-th temporal feature sequence [TeX:] $$d_{i}$$ at time i for the LSTM state [TeX:] $$h_{t}$$ from the target perspective. This approach differs from that of the DARNN model, which uses the input-attention output [TeX:] $$\bar{x}_{i} \text { instead of } d_{i}.$$ As [TeX:] $$\overline{X_{t}}$$ corresponds to the weighted X for selection of important stock and is distorted along the time axis, it is unsuitable as input for learning the stock change pattern of the original sequence X. Therefore, we use [TeX:] $$d_{i},$$ which is the temporal feature sequence produced from X, to produce [TeX:] $$r_{t}^{i}$$ in our proposed method. As we consider the time and stock axes to obtain [TeX:] $$u_{t}, \text { use of } u_{t}$$ overcomes the long-term dependency problem, because stock prices are weighted according to the importance of each data of sequence regardless of distance. In addition, [TeX:] $$\mathbf{u}_{t}$$ overcomes the stock volatility delay problem by selecting important times without using the temporally distorted input-attention output. 3.4 LSTM of 2D-ALSTMAfter [TeX:] $$u_{t}$$ is produced by the 2D-attention component, it is concatenated with the stock index at time t, [TeX:] $$$$ i.e., [TeX:] $$y_{t},$$ and used as the LSTM input to produce the 2D-ALSTM output [TeX:] $$h_{t}.$$ Therefore, for a given X and y, 2D-ALSTM uses the CNN, input attention mechanism, temporal attention mechanism, and LSTM to produce [TeX:] $$h_{t}.$$ We can summarize the 2D-ALSTM procedure as follows:
(23)[TeX:] $$\bar{X}_{t}=\text { input-attention }\left(\boldsymbol{h}_{t-2}, \boldsymbol{h}_{t-1}, \boldsymbol{d}^{k}, \boldsymbol{X}\right)$$
(24)[TeX:] $$\boldsymbol{u}_{t}=\text { temporal-attention }\left(\boldsymbol{h}_{t-2}, \boldsymbol{h}_{t-1}, \bar{X}_{t}, \boldsymbol{d}_{i}\right)$$
(25)[TeX:] $$\boldsymbol{h}_{t}=f_{\text {lstm}}\left(h_{t-1},\left[\boldsymbol{u}_{t} ; y_{t}\right]\right)$$
(26)[TeX:] $$\boldsymbol{H}=\left(\boldsymbol{h}_{\mathbf{1}}, \boldsymbol{h}_{2}, \ldots, \boldsymbol{h}_{\boldsymbol{T}}\right)$$We define a function 2D-ALSTM to represent the 2D_ALSTM described by Eqs. (22)–(26). As the parameter input [TeX:] $$X \in \mathbb{R}^{n \times T}$$ and the output [TeX:] $$\mathbf{H} \in \mathbb{R}^{p \times T}$$ have the same length, i.e., T, the following multilayer structure is possible:
where L is the number of layers. In the experiments performed in this study, we used three 2D-ALSTM layers, with the last cell state [TeX:] $$$$ [TeX:] $$\boldsymbol{h}_{T}^{3}$$ of the last layer [TeX:] $$H^{3}$$ being taken as the linear regression layer input for the final prediction [TeX:] $$\bar{y}_{T+1}:$$
where [TeX:] $$W_{y} \in \mathbb{R}^{1 \times \mathrm{p}} \text { and } b_{y} \in \mathbb{R}^{1}$$ are weight parameters to be learned. 3.5 Loss FunctionIn our proposed method, we use the root mean-squared error (RMSE) as the loss function for training. The loss function for the mini batch [TeX:] $$\mathrm{M}=\left(X^{i}, y^{i}\right), 1 \leq i \leq N$$ can be defined as follows:
(30)[TeX:] $$\text { loss }=\sqrt{\frac{1}{N} \sum_{i=1}^{N}\left(\bar{y}_{T+1}^{i}-y_{T+1}^{i}\right)^{2}}$$We use the Adam optimizer for the model weight parameter training. 4. ExperimentsIn this section, we describe the experiments performed for comparison of the 2D-ALSTM model performance with those of existing models. First, we describe the dataset used for model training and evaluation. Second, we describe the parameter settings of the 2D-ALSTM model and the other baseline models. Third, we report and discuss the experiment results. All experiments were conducted using the TensorFlow deep learning framework, and an NVIDIA GeForce GTX 1070 GPU was used to speed up the model training. 4.1 DatasetsWe used two datasets, KOSPI100 and NASDAQ100, as described in Table 1. Table 1.
The first dataset was the KOSPI100 dataset obtained from the KOSPI stock market. The KOSPI100 dataset consists of the KOSPI100 index and the 100 stock prices used for index calculation. The stock closing prices were used as the model stock price sequence and the index closing prices were used as the model index sequence. The dataset frequency was day-by-day. The index and stock closing prices spanned the period from June 10, 2010, to July 16, 2018, corresponding to a total of 2000 open market days. Only stock for which all data were available for the full 2000 days were selected; thus, a total of 85 stock were selected from the original 100. Samples were created using the obtained 2000-day stock and index data. Each sample incorporated input sequences and a target index value for model training and evaluation. For a fixed time window of size T = 10 and a stride of 1, we generated 1990 samples from the KOSPI100 dataset. We used 1791 samples for training and the remaining 199 samples for evaluation (Table 1). The second dataset was the NASDAQ100 dataset obtained from the NASDAQ stock market. The NASDAQ100 dataset consists of the NASDAQ100 index and the 100 stock prices used for index calculation. Note that this identical dataset was also employed in the study in which DARNN was proposed [10]. As for the KOSPI100 dataset, the stock and index closing prices were used as the stock price sequence and index sequence of the model, respectively. The dataset frequency was minute-byminute. The index and stock closing prices spanned the period from July 26, 2016, to December 22, 2016, corresponding to a total of 105 days. The data for each day contained 390 index and 390 stock prices, excluding November 25 and December 22. Samples were created using the obtained 50560-minute-long stock and index datasets. Each sample incorporated input sequences and a target index value for model training and evaluation. For a fixed T = 10 and a stride of 1, we generated 50,550 samples from this NASDAQ100 dataset. We set the training:validation:evaluation ratio to 7:1.5:1.5. Therefore, we used 28378, 6081, and 6091 samples for training, validation, and evaluation, respectively. All data from both datasets were normalized to between 0 and 1 using min-max scaling. 4.2 Parameter SettingsOur 2D-ALSTM can be implemented as a multi-layer structure; thus, the value of L must be decided. In this study, we employed three 2D-ALSTM layers because this structure yielded the best result for the validation samples obtained from the NASDAQ100 dataset. We set the hidden unit size to 24 for all three layers and T to 10, as for DARNN and MI-LSTM. In the experiments of previous studies [9,10] for DARNN and MI-LSTM, T values of 10 yielded the best results. In addition, the parameter settings for these baseline models were selected according to the best experiment results. When the MI-LSTM model was used for stock price prediction, the Pearson correlation coefficient was used to select positively and negatively correlated stocks. Therefore, in our experiments, we employed two models: MI-LSTM_PN, which used the positively and negatively correlated stocks, and MI-LSTM, which did not use those stocks. The numbers of weight parameters for the 2D-ALSTM, DARNN, MI-LSTM_PN, and MI-LSTM models are listed in Table 2. Although our 2D-ALSTM model contained three layers, it had fewer weight parameters than the other models because it had a small hidden unit size of 24 (the hidden unit size was 64 for all other models). 4.3 ResultsWe conducted five experiments. In each experiment, the models were trained and then evaluated using the loss function presented in section 3.5. The experiment results for all models are presented in Table 3. Note that the loss values for the KOSPI100 and NASDAQ100 datasets are equal to the corresponding values listed in Table 3 multiplied by [TeX:] $$10^{-2} \text { and } 10^{-3}$$, respectively. All models were trained and evaluated five times in the same environment. For training with the KOSPI100 dataset, we used a mini batch size of 1791, equivalent to the number of training samples, and we set the learning rate to 0.01. The training was repeated until no further decreases in loss were observed for the test samples, with repetition over 300 epochs. For training with the NASDAQ100 dataset, we used a mini batch size of 1024 and set the learning rate to 0.01. The training was repeated until there were no further decreases in loss for the test samples, with repetition over 30 epochs. As detailed in Table 3 and Fig. 2, the 2D-ALSTM model exhibited the best performance in 9 of 10 experiments. Comparing the average performance of each model for each dataset, 2D-ALSTM exhibited prediction RMSE, MAE (mean absolute error), and MAPE (mean absolute percentage error) increases of 3.83%, 6.64%, and 6.64% and 1.06%, 1.50%, and 1.44% for the KOSPI100 and NASDAQ100 datasets, respectively, compared to the second-best-performing model, DARNN. The MI-LSTM and MILSTM_ PN models exhibited lower performance than the DARNN and 2D-ALSTM models. This seems to have been because MI-LSTM was created for stock price prediction and generates a time step weight for the LSTM output sequence using a relatively simple attention mechanism compared to the other models. Table 3.
In the experiments with the KOSPI100 dataset, the proposed 2D-ALSTM model exhibited approximately 6.64% higher prediction MAE than the DARNN model, as noted above, although the number of weight parameters was only approximately 50% those of the other models. This result shows that the input attention mechanism of the 2D-ALSTM ameliorates the LSTM stock selection problem, and that the temporal attention of the 2D-ALSTM reduces the LSTM long-term dependency and stock volatility delay problems. These problems are not overcome by the MI-LSTM and DARNN models. In the experiments with the NASDAQ100 dataset, however, the 2D-ALSTM model performance was not significantly higher than that of DARNN. We believe this is because the stock volatility delay problem does not seem to arise for the NASDAQ100 dataset, which incorporates minute-by-minute data. 5. ConclusionsThis paper presented a 2D-ALSTM model for stock index prediction, incorporating input attention and temporal attention mechanisms for the weighting of important stocks and important time steps, respectively. The proposed model is designed to overcome the long-term dependency, stock selection, and stock volatility delay problems that affect existing models. This model can not only be used for stock index prediction, but also has the potential to show great performance in other time series prediction tasks. Recently, the attention-based model has achieved positive results in stock index prediction, as the attention mechanism has been suggested in the neural machine translation field. One particular attention-based model is the DARNN, which consists of input attention and temporal attention. The input attention allows model to focus on important stock prices rather than treating all the input stocks prices equally. The subsequent temporal attention allows the model to selectively focus on important times of the stock price pattern, but the temporal attention has the drawback of performance degradation due to the use of distorted temporal information caused by input attention. The proposed 2D-ALSTM model is designed to resolve the problem by using the original data, which are not distorted, to calculate the temporal-attention weight. The 2D-ALSTM model is validated in a comparative experiment with previous models such as DARNN and MI-LSTM, with real stock data being used for training and evaluation, and achieves superior performance relative to other attention-based models for stock index prediction when applied to a KOSPI100 dataset. BiographyBiographyYoon-Joong Kimhttps://orcid.org/0000-0002-5451-5558He received a B.S. degree in electronics engineering in 1981, an M.S. degree in computer engineering in 1983, and a Ph.D. degree in computer engineering in 1999 from Choongnam National University, Daejeon, Korea. He is a member of recognized scientific institutions such as the IEEE Computational Intelligence Society, the Korean Institute of Information Scientists and Engineers, Korea Information Processing Society, the Korean Institute of Communications and Information Sciences. Since 1984, he has been a full professor in the Department of Computer Engineering of Hanbat National University. He has conducted more than 20 government related projects in his field of expertise. He is an author of more than 30 articles in journals and conference proceedings. His research interests include artificial intelligence, deep learning, and speech recognition. References
|