A Financial Embedded Vector Model and Its Applications to Time Series Forecasting
Keywords:
Embedded Vector, Financial Daily Vector, Financial Weekly Vector, RBF Neural NetworkAbstract
Inspired by the embedding representation in Natural Language Processing (NLP), we develop a financial embedded vector representation model to abstract the temporal characteristics of financial time series. Original financial features are discretized firstly, and then each set of discretized features is considered as a “word” of NLP, while the whole financial time series corresponds to the “sentence” or “paragraph”. Therefore the embedded vector models in NLP could be applied to the financial time series. To test the proposed model, we use RBF neural networks as regression model to predict financial series by comparing the financial embedding vectors as input with the original features. Numerical results show that the prediction accuracy of the test data is improved for about 4-6 orders of magnitude, meaning that the financial embedded vector has a strong generalization ability.References
Scheffer M, Carpenter S R, Lenton T M, Bascompte J, Brock W, Dakos V, Van De Koppel J, Van De Leemput I A, Levin S A, Van Nes E H, Pascual M, Vandermeer J. Anticipating Critical Transitions [J]. Science, 2012, 338(6105): 344-348. https://doi.org/10.1126/science.1225244
Cavalcante R C, Brasileiro R C, Souza V L F, Nobrega J P, Oliveira A L I. Computational Intelligence and Financial Markets: A Survey and Future Directions [J]. Expert Systems with Applications, 2016, 55: 194-211. https://doi.org/10.1016/j.eswa.2016.02.006
Tsai C F, Hsiao Y C. Combining multiple feature selection methods for stock prediction: Union, intersection, and multi-intersection approaches [J]. Decision Support Systems, 2010, 50(1): 258-269. https://doi.org/10.1016/j.dss.2010.08.028
Jasemi M, Kimiagari A M, Memariani A. A modern neural network model to do stock market timing on the basis of the ancient investment technique of Japanese Candlestick [J]. Expert Systems with Applications, 2011, 38(4): 3884-3890. https://doi.org/10.1016/j.eswa.2010.09.049
Shen W, Guo X, Wu C, Wu D. Forecasting stock indices using radial basis function neural networks optimized by artificial fish swarm algorithm [J]. Knowledge-Based Systems, 2011, 24(3): 378-385. https://doi.org/10.1016/j.knosys.2010.11.001
Xiong T, Bao Y, Hu Z, Chiong R. Forecasting interval time series using a fully complex-valued RBF neural network with DPSO and PSO algorithms [J]. Information Sciences, 2015, 305: 77-92. https://doi.org/10.1016/j.ins.2015.01.029
Akbilgic O, Bozdogan H, Balaban M E. A novel Hybrid RBF Neural Networks model as a forecaster [J]. Statistics and Computing, 2014, 24(3): 365-375. https://doi.org/10.1007/s11222-013-9375-7
Cao L. Support vector machines experts for time series forecasting [J]. Neurocomputing, 2003, 51: 321-339. https://doi.org/10.1016/S0925-2312(02)00577-5
Wang J, Hou R, Wang C, Shen L. Improved v-Support vector regression model based on variable selection and brain storm optimization for stock price forecasting [J]. Applied Soft Computing, 2016, 49: 164-178. https://doi.org/10.1016/j.asoc.2016.07.024
Shen F, Chao J, Zhao J. Forecasting exchange rate using deep belief networks and conjugate gradient method [J]. Neurocomputing, 2015, 167: 243-253. https://doi.org/10.1016/j.neucom.2015.04.071
Kuremoto T, Kimura S, Kobayashi K, Obayashi M. Time series forecasting using a deep belief network with restricted Boltzmann machines [J]. Neurocomputing, 2014, 137: 47-56. https://doi.org/10.1016/j.neucom.2013.03.047
Pulido M, Melin P, Castillo O. Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange [J]. Information Sciences, 2014, 280: 188-204. https://doi.org/10.1016/j.ins.2014.05.006
Lecun Y, Bengio Y, Hinton G. Deep learning [J]. Nature, 2015, 521(7553): 436-444. https://doi.org/10.1038/nature14539
Al-Ayyoub M, Nuseir A, Alsmearat K, Jararweh Y, Gupta B. Deep learning for Arabic NLP: A survey [J]. Journal of Computational Science.
Khan W, Daud A, Nasir A A, Amjad T. A survey on the state-of-the-art machine learning models in the context of NLP [J]. Kuwait Journal of Science, 2016, 43(4): 95-113.
Sun S, Luo C, Chen J. A review of natural language processing techniques for opinion mining systems [J]. Information Fusion, 2017, 36(Supplement C): 10-25. https://doi.org/10.1016/j.inffus.2016.10.004
Turian J, Ratinov L, Bengio Y. Word representations: a simple and general method for semi-supervised learning[A]. In proceedings of the 48th Annual Meeting of the Association for Computational Linguistics[C]. Uppsala, Sweden: Association for Computational Linguistics, 2010:384-394
Deerwester S, Dumais S T, Furnas G W, Landauer T K, Harshman R. Indexing by Latent Semantic Analysis. Journal of the American Society for Information Science, 1990, 41: 391-407. https://doi.org/10.1002/(SICI)1097-4571(199009)41:63.0.CO;2-9
Blei D M, Ng A Y, Jordan M I. Latent Dirichlet Allocation. Journal of Machine Learning Research [J]. 2003, 3 (4-5): 993-1022.
Bengio Y, Ducharme J, Vincent P, Janvin C. A neural probabilistic language model [J]. The Journal of Machine Learning Research, 2003, 3(2): 1137-1155.
Mikolov T, Chen K, Corrado G, Dean J. Efficient Estimation of Word Representations in Vector Space[J]. In proceedings of 1st International Conference on Learning Representations (ICLR2013). Scottsdale, AZ, USA, 2013.
Mikolov T, Sutskever I, Chen K, Corrado G, Dean J. Distributed Representations of Words and Phrases and their Compositionality[A]. In proceedings of the 26th International Conference on Neural Information Processing Systems[C]. Lake Tahoe, Nevada: Curran Associates Inc., 2013:3111-3119.
Le Q V, Mikolov T. Distributed Representations of Sentences and Documents[A]. In proceedings of the 31st International Conference on Machine Learning[C]. Beijing, China, 2014:1188-1196.
Singh S P, Kumar A, Darbari H, Singh L, Rastogi A, Jain S. Machine translation using deep learning: An overview[A]. In proceedings of 2017 International Conference on Computer, Communications and Electronics (Comptelix)[C], 2017:162-167.
Frome A, Corrado G, Shlens J, Bengio S, Dean J, Ranzato M A, Mikolov T. DeViSE: A Deep Visual-Semantic Embedding Model[A]. In proceedings of Advances in Neural Information Processing Systems 26 (NIPS 2013)[C]. Lake Tahoe, Nevada, USA: Curran Associates, Inc., 2013:2121-2129.
Moyano L G. Learning network representations [J]. The European Physical Journal Special Topics, 2017, 226(3): 499-518. https://doi.org/10.1140/epjst/e2016-60266-2
Perozzi B, Al-Rfou R, Skiena S. DeepWalk: online learning of social representations[A]. In proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining[C]. New York, USA: ACM, 2014:701-710.
Tang J, Qu M, Wang M, Zhang M, Yan J, Mei Q. LINE: Large-scale Information Network Embedding[A]. In proceedings of the 24th International Conference on World Wide Web[C]. Florence, Italy: International World Wide Web Conferences Steering Committee, 2015:1067-1077.
Yahoo finance [DB/OL]. http://finance.yahoo.com, 2017, 1,1.
Published
Issue
Section
License
ONLINE OPEN ACCES: Acces to full text of each article and each issue are allowed for free in respect of Attribution-NonCommercial 4.0 International (CC BY-NC 4.0.
You are free to:
-Share: copy and redistribute the material in any medium or format;
-Adapt: remix, transform, and build upon the material.
The licensor cannot revoke these freedoms as long as you follow the license terms.
DISCLAIMER: The author(s) of each article appearing in International Journal of Computers Communications & Control is/are solely responsible for the content thereof; the publication of an article shall not constitute or be deemed to constitute any representation by the Editors or Agora University Press that the data presented therein are original, correct or sufficient to support the conclusions reached or that the experiment design or methodology is adequate.