Using Kernel Method to Include Firm Correlation for Stock Price Prediction (2024)

  • Journal List
  • Comput Intell Neurosci
  • v.2022; 2022
  • PMC9005277

As a library, NLM provides access to scientific literature. Inclusion in an NLM database does not imply endorsem*nt of, or agreement with, the contents by NLM or the National Institutes of Health.
Learn more: PMC Disclaimer | PMC Copyright Notice

Using Kernel Method to Include Firm Correlation for Stock Price Prediction (1)

Computational Intelligence and Neuroscience

Comput Intell Neurosci. 2022; 2022: 4964394.

Published online 2022 Apr 5. doi:10.1155/2022/4964394

PMCID: PMC9005277

PMID: 35422853

Hang XuUsing Kernel Method to Include Firm Correlation for Stock Price Prediction (2)

Author information Article notes Copyright and License information PMC Disclaimer

Associated Data

Data Availability Statement

Abstract

In this work, we propose AGKN (attention-based graph learning kernel network), a novel framework to incorporate information of correlated firms of a target stock for its price prediction in an end-to-end way. We first construct a stock-axis attention module to extract dynamic and asymmetric spatial correlations through the kernel method and a graph learning module into which more accurate information can be integrated. An ensemble time-axis attention module is then applied to learn temporal correlations within each stock and market index. Finally, we utilize a transformer encoder to jointly attend to obtain information from different levels for correlations' aggregation and prediction. Experiments with data collected from the Chinese stock market show that AGKN outperforms state-of-the-art baseline methods, making up to 4.3% lower error than the best competitors. The ablation study shows that AGKN pays more attention to hidden correlation between stocks, which improves model's performance greatly.

1. Introduction

Stock prediction has drawn a lot of attention from researchers for a long time. Accurate stock prediction can help investors make appropriate decisions, acting as a meaningful index of portfolio allocation [1]. Efficient market hypothesis (EMH) assumes that all the participants in the market are informationally efficient and that all deals are traded at fair values [2], which mean that the stock market can respond to new information rapidly and stock prices are full representations of market. Therefore, it is impossible to predict stock prices under EMH's framework. However, the rationality of this framework is questionable in reality. Behavioral economists then propose adaptive market hypothesis (AMH) to revise EMH [3]. AMH takes interaction between market and participants into consideration and indicates that it is feasible to obtain excess return through information asymmetry, which implies that the stock is predictable.

Traditional time-series forecasting methods include the AR model, the ARMA model, and the ARIMA model which take single time series as input and make prediction through fitting. Those models can only extract the linear relationship of time series, but the correlation in financial data is always nonlinear [4]. Recently, machine learning and deep learning have become gradually popular in stock prediction. Lee et al. use SVM to extract features and make prediction [5]. Ballings et al. forecast stock price based on ensembles' algorithm including random forest, Adaboost, and kernel factory [6]. Such methods for individual prediction focus on temporal prices pattern of each stock, ignoring the interaction between stocks.

Some previous works utilize models based on LSTM and Markov random fields (MRF) for correlated stocks prediction [7, 8]. Qin et al. propose dual-stage attention-based recurrent neural network (DA-RNN) to learn relationships between stocks automatically [9]. Yoo et al. propose data-axis transformer with the multilevel context (DTML) model for stock movement prediction [8]. DTML utilizes attention-based LSTM and multihead attention for time-axis and data-axis relation extraction, respectively. However, the data-axis module follows time axis, which means that time-axis attention will influence model performance greatly. This will lead to unstable results and poor generalization ability of the model. One of the key challenges for multiple stocks prediction is “how to correlate historical price data both in stock axis and time axis”. Many researchers choose to use graphing neural networks (GNNs) to address this issue because GNN can effectively extract correlation from nodes [7, 10, 11]. However, GNN has the following limitations. First, it requires graph structure as prior knowledge, which is always not defined in financial data. Second, those predefined structure between stocks is fixed, while true correlations keep changing with market. In order to address these issues, the kernel method-based model has been applied. The kernel method plays an important role for data presentation prior to neural network [12]. After mapping features such as vectors and matrix to high-dimensional space by the kernel method, the well-honed linear model can also achieve high performance. This method has a great advantage: it is nonparametric, which is crucial for computation complex. And it can enhance the expressiveness of raw data and extract useful information from mass data. And using correlated stocks for prediction may introduce too many parameters, as well as information noise. We take advantage of the kernel method to solve those problems.

In this study, we propose an end-to-end framework, AGKN (attention-based graph learning kernel network), for correlated stocks prediction. As demonstrated in Figure 1, our model consists of three parts: stock-axis attention generation module, time-axis attention generation module, and attention aggregation module. In stock axis, correlated stocks' data are projected into an reproducing kernel Hilbert space (RKHS) with the learnable kernel method. A novel adjacent matrix extraction layer is further proposed for stock-axis attention capture based on transformed data. In time-axis, attention-based bidirectional LSTM (A-BiLSTM) is applied to obtain the attention scores. Ultimately, an aggregation module is used for attention scores' aggregation and final prediction. In AGKN, asymmetric and dynamic attention scores are generated by the adjacent matrix extraction layer, avoiding the given disadvantages of GNN, and market data are included to respond to changes in macroenvironment flexibly. The kernel method acts as a data intensifier and significantly improves model performance.

Open in a separate window

Figure 1

The overall architecture of AGKN, which consists of three main modules. First, stock-axis attention module computes the adjacent matrix Mout. Second, several A-BilSTMs extract temporal attention within each stocks and concatenate them as Tout. Finally, transformer encoder is used to aggregate them and make prediction.

Our contributions are summarized as follows:

  1. To our best knowledge, this is the first attempt to study correlated stock prediction based on the kernel method. Because too many features as well as noises exist for correlated stocks' prediction, this nonparametric method can enhance raw data and reduce model parameters to prevent overfitting.

  2. We proposed a novel framework, AGKN, for correlated stocks' prediction in an end-to-end way, which can obtain dynamic data relationships both in stock axis and time axis.

  3. We conducted extensive experiments to evaluate AGKN. Experimental results show that AGKN outperforms the state-of-the-art methods. Furthermore, we analyze the stock-axis attention scores, and it proves that our novel architecture is effective in correlation extraction.

The reminders of this paper are organized as follows. In Section 2, we summarize relevant works in correlated stock prediction and kernel method. In Section 3, we present the main methodologies and components of AGKN. In Section 4, we show the experiment settings, results, and analysis. The conclusion will be given in Section 5.

2. Related Works

2.1. Graph Neural Network and Attention Mechanism for Correlated Stock Prediction

Most stocks can be clustered according to industries [13]. Stocks in the same industry have similar price trends in the long term except short-term random disturbance, which acts as a reliable hint for investors. However it remains a challenge to obtain correlation from historical data. Chen et al. take common shareholders in different stocks as nodes and use graph convolutional network (GCN) to capture correlation, which cannot fully represent real-time market changes due to lags of shareholder information update [11]. Cao et al. combine graph Fourier transformer and discrete Fourier transformer in spectral temporal graph neural network (stem-GNN) and use stem-GNN to model interseries and intraseries correlation [10]. Li et al. encode lead-lag relationships between stocks using weighted higher-order Markov random fields (MRFs) and treat them as nodes of MRFs [7]. However, they all require graph structure as prior knowledge, which is not defined in financial data. In fact, spatial-temporal graph neural networks are probably the most suitable type of GNN for multivariate time series, which takes multivariate time series and an external graph as input [14]. In addition, using graph attention network (GAT) for node-level and semantic-level relation capture has achieved good results [15, 16].

Wu et al. propose graph learning module to automatically extract the unidirected relations among multivariate data. A novel mix-hop propagation layer and dilated inception layer are further applied to capture the spatial and temporal dependencies within each time series [17]. However, because financial data are always high noise, frequent alternations of those modules in the proposed framework may exacerbate noise propagation to reduce the model accuracy.

In order to make full use of correlated stocks data, we propose a novel framework to obtain both temporal and spatial attention from raw data in a synchronized rather than step-by-step way. Because financial data are marked by leptokurtic [18], volatility clustering [19], and low signal-noise ratio (SNR) [20], we need to design dedicated modules for specific tasks based on raw data.

2.2. Kernel Method in Neural Networks

The kernel method is closely related to infinitely wide neural networks. Neal and R.M. point out that a single hidden layer neural network with independent, identically distributed random parameters, in the limit of infinite width, is a function drawn from a Gaussian process (GP) [21]. Recent works extend it to deep neural networks and convolutional neural network (CNN) [22, 23]. Jacot et al. believe that least squares regression based on neural tangent kernel (NTK) equals to infinitely wide neural networks [24]. Lee et al. put forward that optimizing the last layer of an infinite wide neural network is equivalent to a Gaussian process based on neural network [25]. Therefore, the kernel method can be applied as one way to improve the performance of deep neural networks. Mairal et al. propose convolutional kernel network (CKN) to learn to approximate the kernel feature map on training data and present the backpropagation method in CKN, bridging a gap between the kernel method and the neural network [26, 27]. Chen et al. generalize convolutional kernel networks to graph-structured data by representing graphs as a sequence of kernel feature maps. However, none of them focus on financial data, which have distinctive characters.

Liu et al. propose the recurrent convolutional neural kernel (RCNK) model, which learned complementary features from historical data and text data, to predict the stock price movement [28]. RCNK constructs an explicit kernel mapping layer to replace fully connected layers to reduce the number of parameters and maintain accuracy. However, RCKN fails to take correlation between stocks into consideration, which is an important character in financial domain.

Generally, a positive defined kernel implicitly induces reproducing kernel Hilbert space (RKHS) and transforms the linear model in RCKS to nonlinear model in input space. This method can reduce the number of parameters in architecture and enhance data presentation. Therefore, merging kernel method and deep neural network can decrease the computation complexity and the risk of overfitting. Besides, through transformation of the learnable kernel method in network, information hidden in raw data is apt to be extracted in the following modules to improve model accuracy.

3. Proposed Approach

3.1. Overview

AGKN is based on the kernel method and GNN, in which correlated stocks' information is used coherently to improve model performance. As illustrated in Figure 1, AGKN consists of stock-axis attention module, time-axis attention module, and aggregation module. To discover hidden correlation among stocks, our stock-axis attention module utilizes a novel layer to extract adjacency matrix. Prior to it, we use the learnable kernel method, instead of a fixed one, to transform data, acting as a data intensifier. As a result, the data intensifier can adapt itself to uncertain market circ*mstances. The learnable way makes it easy to extract correlation, which addresses the limitations of previous works concluded in Section 2.1. Figure 2 presents the framework of stock-axis attention module. In time axis, attention-based LSTM is used to capture relation within historical data of each stock. Instead of proceeding in lockstep, the operations in both directions are synchronized. It is because financial data are always low-SNR (signal-to-noise ratio) and noise tends to accumulate and propagate in feedforward. The synchronized structure can extract more useful context from raw data and reduce noise generation at the same time, which avoids drawbacks we put forward in Section 2.2. Finally, two attention maps are fed into output module, in which we apply transformer encoder [12, 20] for attention aggregation and a fully connected layer for prediction. In this way, temporal and spatial correlations are effectively combined and AGKN performance advances greatly. In more detail, the core components of our model are illustrated in the following.

Open in a separate window

Ensemble stock-axis attention module. The Gaussian and polynomial kervolution are used to transform data.

3.2. Convolutional Kernel Layer

We first elaborate on the general theory of kernel convolution (kervolution) developed in [26]. The kervolution learns to map the given data into a reproducing kernel Hilbert space (RKHS) with trainable parameters in neural network. In RKHS, we can find a subspace to divide transformed data. Given a set χ, a positive definite kernel K: χ×χ⟶ℝ implicitly defines an RKHS H and a mapping φ: χ⟶ℋ:

K1x,x=φx,φxΗ,

(1)

where 〈., .〉 denote the convolutional operator, x, x′ ∈ χ. Specifically, consider two feature matrix x, x′ ⊂ p∗q, where p and q represent width and height. The kernel method K1 is defined as

K1x,x=xxκ1xx,xx,

(2)

where ‖·‖ denote the usual Euclidean norm and κ1 is a dot product kernel on the sphere, and we require κ1 to be smooth enough and its Taylor expansion have nonnegative coefficients to ensure positive definiteness [12]. Specially, the kernel method can be recursively used to produce multilevel features by turning various pixel attributes into patch level, which is useful for nonlinear representation of neighborhoods [29].

Given a convolution filter ωRk·k and a feature map Ω ⊂ p∗q, x ⊂ Ω, a kervolution is defined as

K1x,ω=xκ1xx,ω.

(3)

Figure 3 illustrates the concrete operation for a single channel.

Open in a separate window

Figure 3

Kervolution for one channel.

Polynomial kervolution is defined as

Kpx,w=xTw+cpdp=j=0dpcpdpjxTwj,

(4)

where w is the convolution filter, x is sliding windows for w, and dpZ+andcpR+ are hyperparameters. Dai et al. show that polynomial kernel performs well for natural language processing (NLP) when dp=2. Wang et al. find dp=3 is more appropriate for matrix recognition [30, 31].

Gaussian kervolution (RBF) is an extension to infinite dimensions:

KGx,w=expα2xw22,

(5)

where ‖·‖2 denote unit ℓ-2 norm and α is a hyperparameter. From Taylor expansion, RBF extends the kernel method to infinite dimensions:

KGx,w=expα2x22+w22i=0xTwii!.

(6)

Additionally, complexity of kervolution is normally O(n) as the same as the inner product of convolution [26]. Therefore, by applying kervolution, AGKN can maintain low computational complexity while focusing on different levels as much as possible.

3.3. Stock-Axis Attention Generation

Some graph-based prediction models follow three main steps [10]. First, historical stock prices are fed into the sequence layer to output node embedding for each stock. These outputs are used to capture short-term price trend. Then, an adjacency matrix is created based on some other open graph sources such as Wiki-data or semantic text data on stock forum. Finally, a graph neural network is employed to combine the node embeddings and adjacency matrix as well as to make prediction. Such predefined and multidata-based models are likely to confuse information hidden in stock market and fail to capture dynamic correlation between stocks in different time steps [32] Those drawbacks restrict effective spatial dependencies learning of models. In this phase, we propose a novel matrix extraction layer to extract dynamic spatial dependencies among nodes (stocks) based on transformed data with the kernel method. Furthermore, influence between stocks is usually asymmetric. For instance, small firms' stock prices tend to be more responsive to changes in other stocks prices than large firm does [33]. Thus, the influence of large firm's stock price on small firm is relatively greater than reverse influence. Our data-driven method addresses these problems, extracting dynamic and asymmetric adjacent matrix automatically, which is illustrated as follows:

M1=tanhEX,

(7)

M2=SoftMaxtanhM1BM1T,

(8)

Mout=αM2+1αP,

(9)

where X denotes feature matrix transformed by kernel method, E, B, and α are learnable parameters, and P represents Pearson correlation coefficient matrix.

Equation (7) is expanded as

M1i,j=tanhj=0mei,jxj,i.

(10)

From this perspective, items in matrix M1 can be viewed as weight-average impact of other stocks and market on specific stock in specific time step. In this way, interactions among nodes are taken into consideration in each time step, which enhances the expressiveness of the model. The asymmetric property of our proposed method is achieved by equation (8), which summarizes information of all interacting nodes through hidden edges in entire rolling windows. B acts as a weight matrix in this equation and SoftMax activation function is employed as regularize of the adjacent matrix.

3.4. Time-Axis Attention Generation

Another important part for price prediction is how to obtain temporal dependencies within historical data. To answer this question, we apply LSTM-based model. LSTM addresses the vanishing gradient problem and has stronger memory within time series. Given feature vectors and state vectors, a LSTM cell computes the current state vectors and feeds them into the next cell. Zhou et al. combined attention mechanism and LSTM to capture different information in a sequence, which can handle time series well so that we introduce it into AGKN [34]. The structure of this model is simple and easy to understand, which is shown below.

Attention LSTM: let H be a matrix consisting of the hidden states in each time-step [h1, h2, ..., hT], where T represents the sequence length, and the time-axis attention is computed as

M=tanhH,α=SoftMaxwTM,Tout=HαT,

(11)

where w is a learnable parameter. The detailed implementation is shown in [34].

3.5. Multilevel Attention Aggregation

Finally, AGKN integrates stock-axis attention and time-axis attention for prediction. In this module, we employ multihead attention to tackle this task for its ability to jointly attend to information from different levels. Given attention matrixes in two directions, Tout and Mout, the input of attention F is defined as

F=MoutTout.

(12)

Thus, multihead attention learns to extract correlation between stocks from hybrid attention matrix F and make final prediction.

Transformer encoder used for final prediction: the process of self-attention is shown as follows:

Q=FWQ,K=FWK,V=FWV,Z0=SoftMaxQKTdkV,

(13)

where dk denotes sequence length of time series and WQ, WK, andWV are learnable parameters.

In order to capture various information, multihead attention defines N sets of WQ, WK, andWV to obtain N attention scores Z1, Z2, ..., ZN. These scores focus on different aspects of data. The attention output is the average of all attention heads. Then, a residual connection is applied to avoid vanishing gradient:

Z=Zaverage+F,

(14)

where Z gives us some hints of correlation between stocks in an asymmetric way. Finally, several fully connected layers and activation functions are used for final prediction.

4. Experiments, Results, and Discussion

In this section, we present our experiment setting and results and give detailed analysis. We compare the performance of AGKN with other baseline models. Furthermore, to verify the importance of the kernel method and the adjacent matrix extraction layer in AGKN, we remove the corresponding modules and find that corresponding models perform poorly from different perspectives. Finally, we analyze the results from stock-axis attention generation to emphasize its function again.

4.1. Dataset

We collect daily data of 13 stocks and 1 market index from Shanghai and Shenzhen stock markets in China from 3 January 2017 to 22 October 2021. The transaction data of each trading day are taken as raw data, including opening price, closing price, the highest price, the lowest price, trading volume, and amount. Table 1 summarizes statistics of dataset. The data are from https://tushare.pro/(access on 1 November 2021).

Table 1

Description of data. The mean and variance of closing, highest, and lowest prices are similar to open, so they are not displayed here. We aim to predict the closing price of 600276.SH. Market index is 930767.CSI.

CodeOpenVolumeAmount
MeanStdMean [× 104]Std [× 104]Mean [× 105]Std [× 105]
600276.SH73.3416.5922.0679515.5135416.5741911.52752
600763.SH117.45102.493.6069542.1588864.2677534.771402
300122.SZ73.357.0612.570517.28109611.0571912.44737
600085.SH30.373.710.007569.4407573.1437933.304401
000423.SZ45.6412.326.3193083.8505512.9280012.047151
601607.SH20.972.7114.7790212.05023.1153762.643739
300003.SZ28.446.4718.7386111.654055.5660673.819116
002038.SZ21.3110.088.791727.8224091.7603331.613011
300244.SZ26.967.397.2839115.0862872.1103581.930123
000963.SZ40.7517.9519.1033220.315886.6980948.077395
000931.SZ7.791.3111.2813212.795150.9580111.216137
600196.SH37.9513.3432.1214128.1060514.3109817.73849
603883.SH63.1413.811.9729231.3288421.323861.098693
930767.CSI5188.66773.3711075.733960.7341621103831315.5

Open in a separate window

We move a lag window with size of h time steps to construct the feature matrix and label it with the next day's closing price. We segment data with the first 70% of the date for the training dataset and final 30% for test data.

4.2. Comparative Models

Firstly, to evaluate the effectiveness of stock and time attention module, we take AGKN without them separately as trial objects. Furthermore, we remove the kernel method and adjacent matrix extraction layer, respectively, and include them in the trial group. In addition, kervolutional neural network (KNN) [31], A-BiLSTM [34], and long- and short-term time-series network (LSTNet) [35] will be used as baseline models for comparative experiments. Table 2 shows the abbreviations of comparative models.

Table 2

Model description.

AbbreviationDescription
AGKNAttention-based graph kernel network
AGKN-KAGKN without the kernel method (K)
AGKN-GAGKN without the graphing learning method(G)
AGKN-SAGKN without stock-axis attention(S)
AGKN-TAGKN without time-axis attention(T)
A-BiLSTMAn attention-LSTM model (state-of-the-art model) [34]
LSTNetA deep neural network, which combines convolutional and recurrent neural network [35]
KNNA model which integrates the kernel method and deep neural networks [31]

Open in a separate window

4.3. Metrics and Hyperparameters

We evaluate the result of stock price prediction by two metrics: mean squared error (MSE) and root mean square percentage error (RMSPE). These two metrics are always used to measure the quality of regression problems, which is defined as follows:

MSE=1Ni=1Nyipreyitrue2,RMSPE=1Ni=1Nyipreyitrueyipre2,

(15)

where ypre is predicted price and ytrue is true price.

After attempting many times, AGKN's hyperparameters are shown as follows. Lag window size h is 20, Gaussian kervolution's kernel size is 4, polynomial kervolution's dp is 4, the number of epochs is 500, the number of A-BiLSTM's layers is 2, and optimizer is Adam optimizer with both learning rate and regularization weight being 0.01.

4.4. Results

We test AGKN and its competitors. The performance is evaluated by MSE and RMSPE. Table 3 shows the performance of the different models, and Figure 4 visualizes the predicted price.

Open in a separate window

Figure 4

Comparison of experiment results. (a) AGKN. (b) AGKN-K. (c) AGKN-G. (d) AGKN-S. (e) AGKN-T. (f) LSTNet. (g) A-BiLSTM. (h) KNN.

Table 3

Experiment results.

ModelMSERMSPE
AGKN17.40420.0480
AGKN-K19.18920.0492
AGKN-G46.82640.0786
AGKN-S19.70960.0580
AGKN-T17.70340.0482
A-BiLSTM44.40000.0915
LSTNet34.39670.0915
KNN1889.53240.5664

Open in a separate window

4.5. Ablation Study

From the results, we can conclude that

  1. Each module improves the model performance, and AGKN with complete structure produces the best results. AGKN outperforms AGKN-G and AGKN-S greatly, which means that G and S are more important in the whole framework. As for AGKN-K, AGKN do not show a huge improvement from it. Actually, as we can see from Figure 4, AGKN and its derivatives all perform bad in the area within red box of AGKN's and AGKN-K's result figures. However, in this area, there is an obvious synergetic effect between predicted and realistic trend for the AGKN model. It means that the kernel method we utilized can make the model pay more attention to hidden information which cannot be extracted with other model. That is what we expected the kernel method to do! AGKN-T's performance does not differ greatly from AGKN. However, its result shows distinctive time lag. In other words, the predicted price is close to its prices of previous few days, which is not what we want. Actually, the time lag is as it should be because T module is used to extract time correlation. Without this module, time-axis information cannot be used effectively, causing the time lag.

  2. AGKN outperforms A-BiLSTM and LSTNet, which means it is far from enough to focus only on temporal correlation within historical data. We need to pay more attention to spatial dependencies. On the contrary, KNN's performance is poor. It is because KNN is of numerical instability. If we apply it alone, we will always fail on training.

  3. Figure 5 shows the concatenated attention scores between stocks in Table 1 with respect to target stock (600276.SH). The scores represent how much target stock considers others at each moment. From the figure, we can observe three main points. First, certain stock scores change smoothly over time. This is natural because stock market is continuous so that the property of a stock does not change instantly at a moment. Second, attention scores vary greatly among stocks both at single time step and in the whole time period. 300122.SZ, 300244.SZ, 000931.SZ, and 603883.SH are especially crucial for target stock. The variety stresses the significance of correlated stocks prediction. Third, attention scores for one stock differ greatly in different periods with large gap. This is because stocks are always volatility clustering and their correlations always change greatly along with time, which corresponds to dynamic character we emphasized above.

    Open in a separate window

    Figure 5

    Concatenated stock-axis attention map for single stock: 600276.SH.

5. Conclusions

In this study, we proposed a hybrid model named attention-based graph learning kernel network (AGKN) for correlated stock prediction. AGKN effectively extracts attention both in time axis and data axis without any prior knowledge. By mapping raw data into RCHKs, we proposed a novel way to exploit the correlations among multiple stock data. The experiment results prove the effectiveness of each section in AGKN. Furthermore, our model outperforms a variety of stock prediction methods and acts as a development of the multivariate stock movement prediction in [8].

Data Availability

The data used to support the findings of this study are available at https://tushare.pro/.

Conflicts of Interest

The author declares that there are no conflicts of interest regarding the publication of this paper.

References

1. Long W., Lu Z., Cui L. Deep learning-based feature engineering for stock price movement prediction. Knowledge-Based Systems. 2019;164(15):163–173. doi:10.1016/j.knosys.2018.10.034. [CrossRef] [Google Scholar]

2. Fama E. F. The behavior of stock-market prices. Journal of Business. 1965;38(1):34–105. doi:10.1086/294743. [CrossRef] [Google Scholar]

3. Lo A. W. Reconciling efficient markets with behavioral finance: the adaptive markets hypothesis. The Journal of Investment Consulting. 2005;7(2):21–44. [Google Scholar]

4. Lütkepohl H. New Introduction to Multiple Time Series Analysis. Berlin/Heidelberg, Germany: Springer Science & Business Media; 2005. [Google Scholar]

5. Lee M.-C. Using support vector machine with a hybrid feature selection method to the stock trend prediction. Expert Systems with Applications. 2009;36(8):10896–10904. doi:10.1016/j.eswa.2009.02.038. [CrossRef] [Google Scholar]

6. Ballings M., Van den Poel D., Hespeels N., Gryp R. Evaluating multiple classifiers for stock price direction prediction. Expert Systems with Applications. 2015;42(20):7046–7056. doi:10.1016/j.eswa.2015.05.013. [CrossRef] [Google Scholar]

7. Li C., Song D., Tao D. Multi-task recurrent neural networks and higher-order Markov random fields for stock price movement prediction: multi-task RNN and higer-order MRFs for stock price classification. Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining; 2019 July; New York, NY, USA. pp. 1141–1151. [Google Scholar]

8. Yoo J., Soun Y., Park Y. C., Kang U. Accurate multivariate stock movement prediction via data-Axis transformer with multi-level contexts. Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining; 2021 August; Singapore. pp. 2037–2045. [CrossRef] [Google Scholar]

9. Qin Y., Song D., Chen H., Cheng W., Jiang G., Cottrell G. A dual-stage attention-based recurrent neural network for time series prediction. 2017. https://arxiv.org/abs/1704.02971.

10. Cao D., Wang Y., Duan J., et al. Spectral temporal graph neural network for multivariate time-series forecasting. 2021. https://arxiv.org/abs/2103.07719.

11. Chen Y., Wei Z., Huang X. Incorporating corporation relationship via graph convolutional neural networks for stock price prediction. Proceedings of the 27th ACM International Conference on Information and Knowledge Management; 2018, October; Torino Italy. pp. 1655–1658. [Google Scholar]

12. Scholkopf B., Smola A. J. Adaptive Computation and Machine Learning Series. Cambridge, MA, USA: MIT Press; 2018. Learning with kernels: support vector machines, regularization, optimization, and beyond. [Google Scholar]

13. Bini B. S., Mathew T. Clustering and regression techniques for stock prediction. Procedia Technology. 2016;24:1248–1255. doi:10.1016/j.protcy.2016.05.104. [CrossRef] [Google Scholar]

14. Li M., Zhu Z. Spatial-temporal fusion graph neural networks for traffic flow forecasting. 2020. https://arxiv.org/abs/2012.09641.

15. Wang X., Ji H., Shi C., et al. Heterogeneous graph attention network. Proceedings of the The World Wide Web Conference; 2019, May; San Francisco, CA, USA. pp. 2022–2032. [CrossRef] [Google Scholar]

16. Veličković P., Cucurull G., Casanova A., Romero A., Lio P., Bengio Y. Graph attention networks. 2017. https://arxiv.org/abs/1710.10903.

17. Wu Z., Pan S., Long G., Jiang J., Chang X., Zhang C. Connecting the dots: multivariate time series forecasting with graph neural networks. Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining; 2020, August; New York, NY, USA. pp. 753–763. [CrossRef] [Google Scholar]

18. Mandelbrot B. B. Fractals and Scaling in Finance. New York, NY, USA: Springer; 1997. The variation of certain speculative prices; pp. 371–418. [CrossRef] [Google Scholar]

19. Lux T., Marchesi M. Volatility clustering in financial markets: a microsimulation of interacting agents. International Journal of Theoretical and Applied Finance. 2000;3(4):675–702. doi:10.1142/s0219024900000826. [CrossRef] [Google Scholar]

20. De Long J. B., Shleifer A., Summers L. H., Waldmann R. J. Noise trader risk in financial markets. Journal of Political Economy. 1990;98(4):703–738. doi:10.1086/261703. [CrossRef] [Google Scholar]

21. Neal R. M. Bayesian Learning for Neural Networks. New York, NY, USA: Springer; 1996. Priors for infinite networks; pp. 29–53. [CrossRef] [Google Scholar]

22. Matthews A. G. D. G., Rowland M., Hron J., Turner R. E., Ghahramani Z. Gaussian process behaviour in wide deep neural networks. 2018. https://arxiv.org/abs/1804.11271.

23. Novak R., Xiao L., Lee J., et al. Bayesian deep convolutional networks with many channels are Gaussian processes. 2018. https://arxiv.org/abs/1810.05148.

24. Jacot A., Gabriel F., Hongler C. Neural tangent kernel: convergence and generalization in neural networks. 2018. https://arxiv.org/abs/1806.07572.

25. Lee J., Bahri Y., Novak R., Schoenholz S. S., Pennington J., Sohl-Dickstein J. Deep neural networks as Gaussian processes. 2017. https://arxiv.org/abs/1711.00165.

26. Mairal J., Koniusz P., Harchaoui Z., Schmid C. Convolutional kernel networks. Advances in Neural Information Processing Systems. 2014;27:2627–2635. [Google Scholar]

27. Mairal J. End-to-end kernel learning with supervised convolutional kernel networks. Advances in Neural Information Processing Systems. 2016;29:1399–1407. [Google Scholar]

28. Liu S., Zhang X., Wang Y., Feng G. Recurrent convolutional neural kernel model for stock price movement prediction. PLoS One. 2020;15(6) doi:10.1371/journal.pone.0234206.e0234206 [PMC free article] [PubMed] [CrossRef] [Google Scholar]

29. Bo L., Lai K., Ren X., Fox D. Object recognition with hierarchical kernel descriptors. Proceedings of the Computer Vision & Pattern Recognition; June 2011; Colorado Springs, CO, USA. IEEE; [CrossRef] [Google Scholar]

30. Dai J., Qi H., Xiong Y., et al. IEEE; 2017. Deformable convolutional networks. https://arxiv.org/abs/1703.06211. [Google Scholar]

31. Wang C., Yang J., Xie L., Yuan J. Kervolutional neural networks. 2019. https://arxiv.org/abs/1904.03955.

32. Silvennoinen A., Thorp S. Financialization, crisis and commodity correlation dynamics. Journal of International Financial Markets, Institutions and Money. 2013;24:42–65. doi:10.1016/j.intfin.2012.11.007. [CrossRef] [Google Scholar]

33. Cheung Y.-W., Ng L. K. Stock price dynamics and firm size: an empirical investigation. The Journal of Finance. 1992;47(5):1985–1997. doi:10.1111/j.1540-6261.1992.tb04693.x. [CrossRef] [Google Scholar]

34. Zhou P., Shi W., Tian J., Qi Z., Xu B. Attention-based bidirectional long short-term memory networks for relation classification. Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics; August 2016; Berlin, Germany. [CrossRef] [Google Scholar]

35. Lai G., Chang W. C., Yang Y., Liu H. Modeling long-and short-term temporal patterns with deep neural networks. Proceedings of the the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval; 2018, June; Ann Arbor, MI, USA. pp. 95–104. [Google Scholar]

Articles from Computational Intelligence and Neuroscience are provided here courtesy of Hindawi Limited

Using Kernel Method to Include Firm Correlation for Stock Price Prediction (2024)

FAQs

Which methods is best used for predicting the price of a stock? ›

Predicting stock price with Moving Average (MA) technique. MA is a popular method to smooth out random movements in the stock market.

What is the best algorithm for predicting stock prices? ›

The LSTM algorithm has the ability to store historical information and is widely used in stock price prediction (Heaton et al. 2016).

What is the most accurate stock predictor model? ›

1. AltIndex – Overall Most Accurate Stock Predictor with Claimed 72% Win Rate. From our research, AltIndex is the most accurate stock predictor to consider today. Unlike other predictor services, AltIndex doesn't rely on manual research or analysis.

What is the best AI model for stock prediction? ›

AI-based high-frequency trading (HFT) emerges as the undisputed champion for accurately predicting stock prices. The AI algorithms execute trades within milliseconds, allowing investors and financial institutions to capitalize on minuscule price discrepancies.

What is the formula for predicting the future price of a stock? ›

The formula is shown above (P/E x EPS = Price). According to this formula, if we can accurately predict a stock's future P/E and EPS, we will know its accurate future price. We use this formula day-in day-out to compute financial ratios of stocks.

What are the mathematical methods to predict stock prices? ›

The P/E multiple or price/earnings ratio compares the closing price of the stock with the earnings of the last 12 months. A high value is often a reflection of lofty expectations of stock price and may indicate that the stock is overpriced.

What is the best website for stock analysis? ›

  1. Best overall: Stock Analysis. ...
  2. Best for opinionated research: Seeking Alpha. ...
  3. Best for charts and technical analysis: TradingView. ...
  4. Best for paid stock recommendations: Motley Fool. ...
  5. Best for mutual funds: Morningstar. ...
  6. Best Bloomberg terminal alternative: Koyfin. ...
  7. Best for the latest news: Yahoo Finance.
Mar 6, 2024

How accurate is AI stock prediction? ›

The machine learning models can predict stock returns with remarkable accuracy, achieving an average monthly return of up to 2.71% compared to about 1% for traditional methods," adds Professor Azevedo. The study's findings highlight the potential of such technology for the financial market.

Do stock prediction models work? ›

Utilizing a Keras LSTM model to forecast stock trends

At the same time, these models don't need to reach high levels of accuracy because even 60% accuracy can deliver solid returns. One method for predicting stock prices is using a long short-term memory neural network (LSTM) for times series forecasting.

Is there an AI that predicts stocks? ›

We screened 69 titles and read 43 systematic reviews, including more than 379 studies, before retaining 10 for the final dataset. This work revealed that support vector machines (SVM), long short-term memory (LSTM), and artificial neural networks (ANN) are the most popular AI methods for stock market prediction.

Can algorithms predict the stock market? ›

Stock Price Prediction using machine learning algorithm helps you discover the future value of company stock and other financial assets traded on an exchange. The entire idea of predicting stock prices is to gain significant profits. Predicting how the stock market will perform is a hard task to do.

Is there an algorithm for the stock market? ›

Algorithmic Trading Types

They range from simple single-stock to more complex black-box algorithms that analyze market conditions, price moves, and other financial data to execute trades at optimal times for the least cost-to-maximum profit ratio.

Top Articles
Latest Posts
Article information

Author: Virgilio Hermann JD

Last Updated:

Views: 6297

Rating: 4 / 5 (61 voted)

Reviews: 92% of readers found this page helpful

Author information

Name: Virgilio Hermann JD

Birthday: 1997-12-21

Address: 6946 Schoen Cove, Sipesshire, MO 55944

Phone: +3763365785260

Job: Accounting Engineer

Hobby: Web surfing, Rafting, Dowsing, Stand-up comedy, Ghost hunting, Swimming, Amateur radio

Introduction: My name is Virgilio Hermann JD, I am a fine, gifted, beautiful, encouraging, kind, talented, zealous person who loves writing and wants to share my knowledge and understanding with you.