簡易檢索 / 詳目顯示

研究生: 李國成
Kuo-Cheng Li
論文名稱: 結合關鍵詞分析與遞歸神經網路的股價漲跌預測模型
Prediction on Stock Price Variations Using Keyword Analysis and Recurrent Neural Network
指導教授: 呂永和
Yung-Ho Leu
口試委員: 呂永和
Yung-Ho Leu
楊維寧
Wei-Ning Yang
陳雲岫
Yun-Shiow Chen
學位類別: 碩士
Master
系所名稱: 管理學院 - 資訊管理系
Department of Information Management
論文出版年: 2018
畢業學年度: 106
語文別: 中文
論文頁數: 80
中文關鍵詞: 深度學習遞歸神經網路文字探勘股市漲跌預測
外文關鍵詞: Deep Learning, Recurrent Neural Network, Text Mining, Stock Prediction
相關次數: 點閱:305下載:3
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 股票是現今社會中多數人喜歡的投資理財工具,如何選擇適當的買賣時機一直都是投資者關心的議題。目前對於選擇投資標的及買賣操作,多數會採用技術面與基本面分析。但消息面對於股市的影響也非常的大,由於頻繁發佈之股市新聞蘊含大量的市場消息,其內容能夠反映不同層面事件對於股市的影響,可能藉此改變投資人預期心理及買賣策略。因此建構一套能分析個股相關新聞以結合技術面分析之模型,透過深度學習技術判斷買賣時機,提供投資者較為準確的買賣資訊,是本論文研究的主題。本研究針對多家上市公司財務資訊以及新聞資訊進行實驗。主要透過了文字探勘演算法篩選出新聞關鍵詞彙及產生對應重要度分數,進而結合詞向量模型訓練之輸出,針對個股計算每日該個股的新聞向量,並將其結合財務資料以長短期記憶(LSTM)遞歸神經網路建立預測模型,預測股價之漲跌幅。最終比較本研究考慮新聞因素之模型(Text-LSTM)、長短期記憶神經網路(LSTM)以及差分整合移動平均自迴歸模型(ARIMA)的預測準確率。結果顯示Text-LSTM在變動方向性(Directional Symmetry, DS)於個股預測平均結果為61.8%,最高為69.3%,正確率分別超出LSTM及ARIMA模型約6.2%及13.4%。


    Nowadays, the stock is still the most popular investment tool in our society. To find the time of trading for a stock is the most important issue that concerns all the investors. When selecting an investment target and finding the time of trading, most of investors take advantage of technical analysis and fundamental analysis. However, the influence of financial news on the stock market should also be taken into account since the frequently released financial news carry large amount of information that influences the expectation of an investor on a stock. As such, the financial news may affect the trading behavior of the investor. This thesis aims at constructing a prediction model on the variations of a stock price based on the company-related news and technical analysis of the stock. To this end, we collected financial news of several listed stocks in Taiwan Stock Exchange (TWSE). We first use text mining algorithm to filter out important keywords from the news and calculate their corresponding scores of importance. Then, we train to find the representing vectors of the important words using Word2Vec model. Finally, the vectors of important words of the daily news of a stock are multiplied by their corresponding scores of importance to generate a news vector for the stock. With the daily news vector and several financial variables, we construct a prediction model on the selected stocks using the long short-term memory (LSTM) recurrent neural network. We conducted several experiments on the proposed model, termed Text-LSTM, a model without considering the daily news, termed LSTM, and a model constructed using ARIMA model. The experimental results showed that the Text-LSTM achieved an average of 61.8% accuracy and a maximum of 69.3% accuracy in Directional Symmetry (DS) on the predictions. The accuracy of the Text-LSTM outperforms those of the pure LSTM and the ARIMA by 6.2% and 13.4%, respectively.

    摘要 I ABSTRACT II 致謝 III 目錄 IV 圖目錄 VI 表目錄 VIII 第一章 緒論 1 1.1 研究背景 1 1.2 研究動機 2 1.3 研究目的 3 1.4 研究架構 4 第二章 文獻探討 5 2.1 基本面分析 5 2.2 技術面分析 7 2.3 文字探勘 16 2.3.1 TF-IDF 17 2.3.2 Word Embedding 18 2.4 遞歸類神經網路 25 第三章 研究方法 34 3.1 研究架構 34 3.2 資料蒐集 35 3.3 財務資料處理 35 3.3.1 技術指標計算及變數說明 36 3.3.2 變數篩選 37 3.4 文字資料處理 38 3.4.1 中文斷詞 39 3.4.2 TF-IDF計算及關鍵詞過濾 40 3.4.3 詞向量訓練 41 3.5 預測模型建立 47 3.6 模型評估標準 49 3.6.1 評估指標 49 3.6.2 獲利表現 50 第四章 實驗方法與結果分析 51 4.1 實作環境 51 4.2 資料描述與實驗設計 51 4.3 實驗參數設定 53 4.3.1 Word2Vec 模型 53 4.3.2 LSTM 網路模型 53 4.3.3 ARIMA 模型 55 4.4 預測模型結果與評估 55 4.5 統計假設檢定 61 第五章 結論與建議 63 5.1 結論及研究貢獻 63 5.2 未來研究方向 64 參考文獻 65

    [1] Fama, E. F. (1995). Random Walks in Stock Market Prices. Financial Analysts Journal, 51(1), 75-80.
    [2] Keynes, J. M. (1936). The General Theory of Employment, Interest, and Money. New York: Harcourt.
    [3] Mindell, J. (1948). The Stock Market: Basic Guide for Investors. New York: Forbes.
    [4] Graham, B. & Dodd, D. L. (1951). Security analysis: Principles and technique. New York: McGraw-Hill.
    [5] Olson, O. & Mossman, C. (2003). Neural Network Forecasts of Canadian Stock Returns using Accounting Ratios. International Journal of Forecasting, 19(3), 453-465.
    [6] Thawornwong, S. & Enke, D. (2004). The adaptive selection of financial and economic variables for use with artificial neural networks. Neurocomputing, 56, 205-232.
    [7] Chen, Y. J. & Chen, Y. M. (2013). A fundamental analysis-based method for stock market forecasting. Proceedings of 2013 Fourth International Conference on Intelligent Control and Information Processing (ICICIP), 354–359. doi:10.1109/ ICICIP.2013.6568097
    [8] Tsai, C. F. & Hsiao, Y. C. (2010). Combining multiple feature selection methods for stock prediction: Union, intersection, and multi-intersection approaches. Decision Support Systems, 50(1), 258-269.
    [9] Edirisinghe, N.C.P. & Zhang, X. (2007). Generalized DEA model of fundamental analysis and its application to portfolio optimization. Journal of Banking & Finance, 31(11), 3311-3335.
    [10] Lam, M. (2004). Neural network techniques for financial performance prediction: integrating fundamental and technical analysis. Decision Support System, 37(4), 567–581.
    [11] Kirkpatrick II, C. D. & Dahlquist, J. R. (2006). Technical Analysis: The Complete Resource for Financial Market Technicians. FT Press. Retrieved from http://1.droppdf.com/files/wWSus/technical-analysis-the-complete-resource-for-financial-market-technicians-2011.pdf
    [12] Murphy, J. J. (1999). Technical Analysis of the Financial Markets: A Comprehensive Guide to Trading Methods and Applications. New York: New York Institute of Finance.
    [13] 林成蔭(2011)。股海樂活。法意出版社。
    [14] 杜金龍(2002)。技術指標在台灣股市應用的訣竅。財訊出版社。
    [15] Oriani, F. B. & Coelho, G. P. (2016). Evaluating the impact of technical indicators on stock forecasting. Proceedings of 2016 IEEE Symposium Series on Computational Intelligence (SSCI), 1-8. doi:10.1109/SSCI.2016.7850017
    [16] Gholamiangonabadi, D., Mohseni Taheri, S. D., Mohammadi, A. & Menhaj, M. B. (2014). Investigating the performance of technical indicators in electrical industry in Tehran's Stock Exchange using hybrid methods of SRA, PCA and Neural Networks. Proceedings of 2014 5th Conference on Thermal Power Plants (CTPP), 75-82. doi:10.1109/CTPP.2014.7040698
    [17] Ni, L.P., Ni, Z.W. & Gao, Y. Z. (2011). Stock trend prediction based on fractal feature selection and support vector machine. Expert Systems with Applications, 38(5), 5569-5576.
    [18] Chen, J. F., Chen, W. L., Huang, C. P., Huang, S. H. & Chen, A. P. (2016). Financial Time-Series Data Analysis Using Deep Convolutional Neural Networks. Proceedings of 2016 7th International Conference on Cloud Computing and Big Data (CCBD), 87-92. doi:10.1109/CCBD.2016.027
    [19] Bao, W., Yue, J. & Rao, Y. (2017). A deep learning framework for financial time series using stacked autoencoders and long-short term memory. PLOS ONE, 12(7).
    [20] Huang, C. L. & Tsai, C. Y. (2009). A hybrid SOFM-SVR with a filter-based feature selection for stock market forecasting. Expert Systems with Applications, 36(2), 1529-1539.
    [21] Sullivan, D. P. (2001). Document Warehousing and Text Mining. New York: Wiley.
    [22] Salton, G. & McGill, M. J. (1983). Introduction to modern information retrieval. New York: McGraw-HIII.
    [23] Sanchez, N. (2012). Communication Process. Retrieved from
    https://web.njit.edu/~lipuma/352comproc/comproc.htm
    [24] Harris, Z. S. (1954). Distributional structure. Word, 10(2-3), 146-162. doi:10.1080/00437956.1954.11659520
    [25] Firth, J. R. (1957). A synopsis of linguistic theory 1930-55. 1952-59, 1-32.
    [26] Turian, J., Ratinov, L. & Bengio, Y. (2010). Word representations: a simple and general method for semi-supervised learning. Proceedings of the 48th annual meeting of the association for computational linguistics (ACL), 384–394.
    [27] Chen, S. C., Hung, H. T., Chen, B. & Chen, K. Y. (2015). Exploring Word Embedding and Concept Information for Language Model Adaptation in Mandarin Large Vocabulary Continuous Speech Recognition. Proceedings of the 27th Conference on Computational Linguistics and Speech Processing, 4-17.
    [28] Hinton, G. E. (1986). Learning distributed representations of concepts. Proceedings of the Eighth Annual Conference of the Cognitive Science Society, 1–12.
    [29] Mikolov, T., Chen, K., Corrado, G. S., & Dean, J. (2013). Efficient Estimation of Word Representations in Vector Space. CoRR, abs/1301.3781.
    [30] Rumelhart, D.E., Hinton, G. E. & Williams, R. J. (1986). Learning Internal Representation by Error Propagation. Parallel Distributed Processing: Explorations in the Microstructure of Cognition, 1, 318-362.
    [31] Olah, C. (2015). Understanding LSTM Networks. Retrieved from http://colah.github.io/posts/2015-08-Understanding-LSTMs/
    [32] Britz, D. (2015). Recurrent Neural Networks Tutorial: Part 3 – Backpropagation Through Time and Vanishing Gradients. Retrieved from
    http://www.wildml.com/2015/10/recurrent-neural-networks-tutorial-part-3-backpropagation-through-time-and-vanishing-gradients/
    [33] Hochreiter, S. & Schmidhuber, J. (1997). Long short-term memory, Neural computation, 9(8), 1735-1780.
    [34] Friedman, J. H. (1991). Multivariate Adaptive Regression Splines, The Annals of Statistics, 19(1), 1-67.
    [35] Palshikar, G. K. (2007). Keyword extraction from a single document using centrality measures. Proceedings of the International Conference on Pattern Recognition and Machine Intelligence, 503-510.

    QR CODE