簡易檢索 / 詳目顯示

研究生: 姚昌辰
Chang-chen Yao
論文名稱: 以最小平均平方學習法增強貝氏分類器之研究
The Study of a Bayes Classifier Enhanced by Least-Mean-Square Learning Approach
指導教授: 楊英魁
Ying-kuei Yang
口試委員: 孫宗瀛
Tsung-ying Sun
黎碧煌
Bih-hwang Lee
陳俊良
Jiann-liang Chen
學位類別: 碩士
Master
系所名稱: 電資學院 - 電機工程系
Department of Electrical Engineering
論文出版年: 2014
畢業學年度: 102
語文別: 中文
論文頁數: 74
中文關鍵詞: 最小平均平方學習演算法決策樹資料探勘貝氏分類器
外文關鍵詞: least mean square, decision tree, data mining, Bayes classifier
相關次數: 點閱:183下載:0
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 分類在資料探勘中一直是非常重要的一部份,透過資料所提供的資訊有助於人們預測未來可能發生的事件,因此不同的分類器都有其優勢去處理不同特性的分類問題;其中,單純貝氏分類器更是被大家廣為使用的分類器之一,其具有效率快且分類結果穩定等特性,因此在實務上也被廣泛使用。

    本研究以貝氏分類器為基礎,透過決策樹來篩選出重要的屬性,接著利用決策樹的特性來幫忙決定屬性的初始權重(亦即其重要性),然後再使用最小平均平方學習法的方式加以對權重做調整,訓練完成後將貝氏分類器賦予各屬性的權重,用其作為資料的分類和預測。實驗結果顯示針對各種不同性質(亦即不同領域)的資料集有很好的辨識能力,例如糖尿病資料集(醫療)、鳶尾花資料集(植物)、玻璃辨識資料集(物品)等等,並不侷限於某一個特定領域,所以本論文的分類模型具有較佳的強健性。


    Classification is always an important part in data mining. The information of data can help us predicting what will happen in the future. Therefore, there are different merits for different classifiers to process various classified problems. Among many classifiers, Naive Bayes Classifier works more effectively and stably, so it is the most popular classifier in practical applications.

    In this paper, firstly, the Decision Tree based on Bayes Classifier is used to decide the major attributes that are considered to give more contribution in terms of classification. Secondly, the initial weight for each of the selected major attributes is set up based on the nature of Decision Tree in which the level distance from the tree root reflects the role of importance to the classification. Thirdly, the weight of each selected major features is then adjusted by the least mean square learning process. Finally, the resultant weights of major features from the learning are used in the Bayes Classifier for classification and prediction. The experiment shows that the recognizing ability of the classifier proposed in this thesis is relatively better against other some compared approaches for various Datasets in the different domains. The experimental result also shows the proposed approach has better robustness.

    摘要 I Abstract II 致謝 III 目錄 IV 圖目錄 VI 表目錄 VII 第一章緒論 1 1.1研究背景 1 1.2研究動機 3 1.3論文架構 4 第二章文獻探討 5 2.1資料探勘概念 5 2.1.1資料探勘的定義 6 2.1.2資料探勘的流程 8 2.2決策樹 (Decision Tree) 15 2.2.1決策樹之分割方法 17 2.2.2CART決策樹 20 2.2.3 ID3決策樹演算法 21 2.2.4 C4.5決策樹演算法 23 2.3屬性篩選 25 第三章研究方法 28 3.1系統架構 28 3.2 單純貝氏分類器(Naive Bayes Classifier) 29 3.3 最小平均平方演算法(Least Mean Square) 35 3.4基於最小平均平方學習法強化貝氏分類器(Naive Bayes Classifier with Least Mean Square) 36 第四章實驗結果與討論 41 4.1 資料集 41 4.1.1 糖尿病資料集(Pima Indians Diabetes Database) 42 4.1.2 玻璃辨識資料集(Glass Identification Database) 43 4.1.3 鳶尾花資料集(Iris Plants Database) 43 4.1.4影像分割資料集(Image Segmentation Data) 43 4.1.5大黃豆資料集(Large Soybean Database) 43 4.1.6 1984年美國國會投票紀錄資料集(1984 United States Congressional Voting Records Database) 44 4.1.7井字遊戲殘局資料集(Tic-Tac-Toe Endgame Data) 44 4.2 評估方法 45 4.3 實驗結果 47 第五章結論與建議 57 參考文獻 58

    [1] J. Han and M. Kamber, Data Mining: Concepts and Techniques, Morgan Kaufmann Publishers, 2nd edition, 2000.
    [2] M.J.A. Berry and G.S. Linoff, Data Mining Techniques: For Marketing, Sales and Customer Support, Jogn Wiley & Sons, 1997.
    [3] A. Franco-Arcega, J. A. Carrasco-Ochoa, G. Sanchez-Diaz and J. F. Martinez-Trinidad, “Decision tree induction using a fast splitting attribute selection for large datasets,” Expert Systems with Applications, Vol. 38, pp. 14290-14300, 2011.
    [4] B. Aviad and G. Roy, “Classification by clustering decision tree-like classifier based on adjusted clusters,” Expert Systems with Applications, Vol. 38, pp. 8220-8228, 2011.
    [5] L. Breiman, J. H. Friedman, R. A. Olsen and C. J. Stone, “Classification and Regression Trees,” Wadsworth & Brooks, 1984.
    [6] C. E. Shannon, “A Mathematical Theory of Communication,” Reprinted with corrections from The Bell System Technical Journal, Vol. 27, pp. 379–423, 623–656, July, October, 1948.
    [7] J. R. Quinlan, “Induction of decision trees. Machine Learning,” Machine Learning, Vol. 1, pp. 81-106, 1986
    [8] J. R. Quinlan, “Improved Use of Continuous Attributes in C4.5,” Journal of Artificial Intelligence Research, Vol. 4, pp. 77-90, 1996
    [9] J.R. Quinlan, “C4.5: Programs for Machine Learning,” Morgan Kaufmann, 1993
    [10] L. Koc, T.A. Mazzuchi and S. Sarkani, “A network intrusion detection system based on a hidden naive Bayes multiclass classifier,” Expert Systems with Applications, Vol. 39, pp. 13492–13500, 2012.
    [11] M.A. Valle, S. Varas and G.A. Ruz, “ob performance prediction in a call center using a naive Bayes classifier,” Expert Systems with Applications, Vol. 39, pp. 9939–9945, 2012.
    [12] Langley, Pat, Iba, Wayne, Thompson and Kevin, “Analysis of Bayesian classifiers,” Proceedings Tenth National Conference on Artificial Intelligence, pp. 223-228, 1992.
    [13] D. M. Farid and M. Z. Rahman, “Learning intrusion detection based on adaptive bayesian algorithm,” Proceedings of 11th International Conference on Computer and Information Technology, pp. 652-656, 2008.
    [14] K. F. R. Liu and J. S. Chen, “Prediction and assessment of student learning outcomes in calculus a decision support of integrating data mining and Bayesian belief networks,” Computer Research and Development (ICCRD), 2011 3rd International Conference on, Vol. 1, pp. 299-303, 2011.
    [15] J. Koščak, R. Jaksa and P. Sincak, ”tochastic weight update in the backpropagation algorithm on feed-forward neural networks,” Neural Networks (IJCNN), The 2010 International Joint Conference on, pp. 1-4, 2010.

    [16] C. Qing, “The Application of Problems Concerning the Imbalanced Data Classification by Means of Support Vector Machines,” Knowledge Acquisition and Modeling (KAM), 2011 Fourth International Symposium on, pp. 291-295, 2011.
    [17] F. Giannotti, L. V. S. Lakshmanan, A. Monreale, D. Pedreschi and W. Hui, “Privacy-Preserving Mining of Association Rules From Outsourced Transaction Databases,” Systems Journal, IEEE, Vol. 7, pp. 385-395, 2013.
    [18] S. Garcia, J. Derrac, J. R. Cano and F. Herrera, “Prototype Selection for Nearest Neighbor Classification: Taxonomy and Empirical Study,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, Vol. 34, pp. 417-435, 2012.
    [19] T. Quirino, M. Kubat and N. J. Bryan, “Instinct-Based Mating in Genetic Algorithms Applied to the Tuning of 1-NN Classifiers,” Knowledge and Data Engineering, IEEE Transactions on, Vol. 22, pp. 1724-1737, 2010.
    [20] S. Zhao; E. C. C. Tsang, D. Chen; X. Wang, “Building a Rule-Based Classifier—A Fuzzy-Rough Set Approach,” Knowledge and Data Engineering, IEEE Transactions on, Vol. 22, pp. 624-638, 2010.
    [21] D. A. Keim, C. Pansea, M. Sipsa and S. C. Northb, “Pixel based visual data mining of geo-spatial data,” Computers & Graphics, Vol. 28, pp. 327-344, 2004
    [22] S. C. Hui and G. Jha, “Data mining for customer service support,” Information & Management, Vol. 38, pp. 1-13, 2000.

    [23] H. Curt, “The Deville’s in The Detail: Techniques, Tool, and Applications for Data Mining and Knowledge Discovery Part I,” Intelligent Software Strategies, Vol. 6, pp. 3, 1995.
    [24] C. Hall, “The Deville’s in The Detail: Techniques, Tool, and Applications for Data Mining and Knowledge Discovery Part II,” Intelligent Software Strategies, Vol. 6, pp. 1-16, 1995.
    [25] U. M. Fayyad and P. Stolorz, “Data mining and KDD: Promise and challenges,” Further Generation Computer Systems, Vol. 13, pp. 99-115, 1997.
    [26] Data Mining Methodology: http://www.kduggets.com/polls/index.html
    [27] 韋瑞,Data Mining概述:以Clementine 7.0為例,中華資料採礦協會,2003
    [28] R. N. Kristin and I. P. Matkovsky, “Using Data Mining Techniques for Fraud Detection,” SAS Institute Inc. and Federal Data Corporation, 1999.
    [29] Y. L. Zhang, N. Guo, H. Du and W. H. Li, “Automated defect recognition of C-SAM image in IC packaging using support vector machines,” International Journal of Advanced Manufacturing Technology, Vol. 25, pp. 1191 -1196, 2005.
    [30] C. W. Hsu, C. C. Chang and C. J. Lin, “A practical guide to support vector classification,” Department of Computer Science, National Taiwan University, 2003.
    [31] J. Wang, X. Wu and C. Zhang, “Support vector machines based on K-means clustering for real-time business intelligence systems,” International Journal of Business Intelligence and Data Mining, Vol. 1, pp. 54 -64, 2005.
    [32] N. Acır, O. Ozdamar, and C. Guzelis, “Automatic classification of auditory brainstem responses using SVM-based feature selection algorithm for threshold detection,” Engineering Applications of Artificial Intelligence, Vol. 19, pp. 209-218, 2006.
    [33] G. Valentini, M. Muselli and F. Ruffino, “Cancer recognition with bagged ensembles of support vector machines,” Neurocomputing, Vol. 56, pp. 461-466, 2004.
    [34] R. Kohavi and G. H. John, “Wrappers for feature subset selection,” Artificial Intelligence, Vol. 95, pp. 273-324, 1997.
    [35] H. Liu, and H. Motoda, “Feature selection for knowledge discovery and data mining”, Kluwer Academic, 1998.
    [36] D. Mladenic, and M. Grobelnik, “Feature selection on hierarchy of web documents,” Decision Support Systems, Vol. 35, pp. 45 -87, 2003.
    [37] D. M. Farid, L. Zhang, C. M. Rahman, M. A. Hossain, R. Strachan, “Hybrid decision tree and naive Bayes classifiers for multi-class classification tasks,” Expert Systems with Applications, Vol. 41, pp. 1937-1946, 2014.
    [38] 廖述賢、溫志皓,資料探勘理論與應用,博碩文化股份有限公司,2012
    [39] B. Widrow and M. E. Hoff, “Adaptive switching circuits,” IRE WESCON Convention Record, Part 4, pp. 96-104 ,1960.
    [40] W. Gong and X. Wang, “MCMC Bayes-Mixed Logit for corridor transport mode split,” Transportation, Mechanical, and Electrical Engineering (TMEE), 2011 International Conference on, pp. 1867-1871, 2011.
    [41] UCI Machine Learning Repository: http://archive.ics.uci.edu/ml/
    [42] I. Androutsopoulos, J. Koutsias, K. V. Chandrinos, G. Paliouras and C. D. Spyropoulos, "An evaluation of naive bayesian anti-spam filtering," Proc. of the workshop on Machine Learning in the New Information Age, pp. 9-17, 2000.
    [43] R. Kohavi, “A study of cross-validation and bootstrap for accuracy estimation and model selection,” Proceedings of the Fourteenth International Joint Conference on Artificial Intelligence, Vol. 2, pp. 1137-1143, 1995.

    無法下載圖示 全文公開日期 2019/07/28 (校內網路)
    全文公開日期 本全文未授權公開 (校外網路)
    全文公開日期 本全文未授權公開 (國家圖書館:臺灣博碩士論文系統)
    QR CODE