簡易檢索 / 詳目顯示

研究生: 張家瑋
Eldridge Rexroy Edmondo Dealon
論文名稱: 使用數據增強方法改進階層式Transformer在情感分類問題上的效能
Improving the Performance of the Hierarchical Transformer Model in Emotion Classification using Data Augmentat
指導教授: 呂永和
Yung Ho Leu
口試委員: 楊維寧
Wei Ning Yang
陳雲岫
Yun Shiow Chen
學位類別: 碩士
Master
系所名稱: 管理學院 - 資訊管理系
Department of Information Management
論文出版年: 2022
畢業學年度: 110
語文別: 英文
論文頁數: 54
中文關鍵詞: Text ClassificationEmotion ClassificationTransformerEncoderHierarchical TransformerData AugmentationAEDA
外文關鍵詞: Text Classification, Emotion Classification, Transformer, Encoder, Hierararchical Transformer, Data Augmentation, AEDA
相關次數: 點閱:234下載:7
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報

Abstract - Emotion has always been an element of human life, expressed by action or conversation. It can influence how individuals regard one another, and events can influence it in our environment. As a result, it is critical to understand how our surroundings feel emotionally so that we can act appropriately in any situation. people can predict each other emotions by listening to their intonation and the words themselves. However, with text is rather hard to predict emotion since it only has text as an attribute. Therefore, a lot of research has been done to make this task more accurate. This task has been challenging because of its imbalanced class on datasets and contextual understanding between utterance. Existing method rely only rely on hierarchy-based model. In this research, we propose a method of detecting emotion by using a hierarchical transformer with a contextual enhancer. Dataset extreme imbalance class also created a problem to create a good model. Therefore, data augmentation AEDA is also implemented to help and fix the class imbalance. The improvement of the proposed will be shown on EmotionPush and Emory NLP dataset.


Abstract - Emotion has always been an element of human life, expressed by action or conversation. It can influence how individuals regard one another, and events can influence it in our environment. As a result, it is critical to understand how our surroundings feel emotionally so that we can act appropriately in any situation. people can predict each other emotions by listening to their intonation and the words themselves. However, with text is rather hard to predict emotion since it only has text as an attribute. Therefore, a lot of research has been done to make this task more accurate. This task has been challenging because of its imbalanced class on datasets and contextual understanding between utterance. Existing method rely only rely on hierarchy-based model. In this research, we propose a method of detecting emotion by using a hierarchical transformer with a contextual enhancer. Dataset extreme imbalance class also created a problem to create a good model. Therefore, data augmentation AEDA is also implemented to help and fix the class imbalance. The improvement of the proposed will be shown on EmotionPush and Emory NLP dataset.

ABSTRACT i ACKNOWLEDGEMENT ii TABLE OF CONTENTS iii LIST OF FIGURES v LIST OF TABLES vi Chapter 1 1 1.1 BACKGROUND AND MOTIVATION 1 1.2 RESEARCH OBJECTIVE AND LIMITATION 3 1.3 RESEARCH CONTRIBUTIONS 4 1.4 RESEARCH OVERVIEW 4 Chapter 2 5 2.1 RELATED WORKS 5 2.2 BASIC THEORY 5 2.2.1 Machine Learning 6 2.2.2 Artificial Neural Network and Deep Learning 7 2.2.3 Natural Language Processing (NLP) 10 2.3 DATA AUGMENTATION 11 2.3.1 AEDA 11 2.4 TRANSFORMER 12 2.4.1 Self-Attention 13 2.4.2 Encoder 18 2.4.3 Decoder 18 2.5 BERT 19 2.6 COMPARED METHOD 21 Chapter 3 24 3.1 LITERATURE STUDY 24 3.2 DATASETS 24 3.3 METHOD AND ARCHITECTURE 25 3.3.1 Individual Utterance Embedding 26 3.3.2 Contextual Utterance Embedding 26 3.3.3 Enhanced Contextual Utterance and Fully Connected 28 3.3.4 Architecture Differences from Hierarchical Transformer Model 29 Chapter 4 30 4.1 EXPERIMENTAL SETTINGS 30 4.1.1 Evaluation Metrics 30 4.1.2 Dataset 30 4.1.3 Loss Function 31 4.1.4 Experiment Parameters and Details 31 4.1.5 Data Augmentation Addition 33 4.1.6 Experiment Result 34 4.1.7 Result Comparison 38 4.1.8 Overall View 40 Chapter 5 41 5.1 CONCLUSION 41 5.2 FUTURE RESEARCH 42 Reference 43

[1] A. Seyeditabari, N. Tabari, S. Gholizadeh, and W. Zadrozny, "Emotion Detection in Text: Focusing on Latent Representation," Jul. 2019, [Online]. Available: http://arxiv.org/abs/1907.09369
[2] P. Ekman, "An argument for basic emotions," Cognition and Emotion, vol. 6, no. 3–4, pp. 169–200, May 1992, doi: 10.1080/02699939208411068.
[3] W. Gerrod Parrott, Emotions in Social Psychology: Essential Readings, Illustrated. 2001.
[4] L. Feldman Barrett and J. A. Russell, "Independence and bipolarity in the structure of current affect.," Journal of Personality and Social Psychology, vol. 74, no. 4, pp. 967–984, 1998, doi: 10.1037/0022-3514.74.4.967.
[5] R. PLUTCHIK, "Chapter 1 - A GENERAL PSYCHOEVOLUTIONARY THEORY OF EMOTION," in Theories of Emotion, R. Plutchik and H. Kellerman, Eds. Academic Press, 1980, pp. 3–33. doi: https://doi.org/10.1016/B978-0-12-558701-3.50007-7.
[6] A. Vaswani et al., "Attention Is All You Need," Jun. 2017, [Online]. Available: http://arxiv.org/abs/1706.03762
[7] J. Devlin, M.-W. Chang, K. Lee, K. T. Google, and A. I. Language, "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding." [Online]. Available: https://github.com/tensorflow/tensor2tensor
[8] Y. Liu et al., "RoBERTa: A Robustly Optimized BERT Pretraining Approach," Jul. 2019, [Online]. Available: http://arxiv.org/abs/1907.11692
[9] Q. Li, C. Wu, Z. Wang, and K. Zheng, "Hierarchical transformer network for utterance-level emotion recognition," Applied Sciences (Switzerland), vol. 10, no. 13, Jul. 2020, doi: 10.3390/app10134447.
[10] L. Luo, H. Seng, H. Yang, and F. Y. L. Chin, "EmotionX-DLC: Self-Attentive BiLSTM for Detecting Sequential Emotions in Dialogues," 2018.
[11] W. Jiao, H. Yang, I. King, and M. R. Lyu, "HiGRU: Hierarchical Gated Recurrent Units for Utterance-level Emotion Recognition," Apr. 2019, [Online]. Available: http://arxiv.org/abs/1904.04446
[12] Y.-H. Huang, S.-R. Lee, M.-Y. Ma, Y.-H. Chen, Y.-W. Yu, and Y.-S. Chen, "EmotionX-IDEA: Emotion BERT -- an Affectional Model for Conversation," Aug. 2019, [Online]. Available: http://arxiv.org/abs/1908.06264
[13] J. Lee and W. Lee, "CoMPM: Context Modeling with Speaker's Pre-trained Memory Tracking for Emotion Recognition in Conversation," Aug. 2021, [Online]. Available: http://arxiv.org/abs/2108.11626
[14] J. Li, M. Zhang, D. Ji, and Y. Liu, "Multi-Task Learning with Auxiliary Speaker Identification for Conversational Emotion Recognition," Mar. 2020, [Online]. Available: http://arxiv.org/abs/2003.01478
[15] O. Simeone, "A Very Brief Introduction to Machine Learning With Applications to Communication Systems," Aug. 2018, [Online]. Available: http://arxiv.org/abs/1808.02342
[16] K. (Kevin N. ) Gurney, An introduction to neural networks. UCL Press, 1997.
[17] C. Janiesch, P. Zschech, and K. Heinrich, “Machine learning and deep learning,” Electronic Markets, vol. 31, no. 3, pp. 685–695, Sep. 2021, doi: 10.1007/s12525-021-00475-2.
[18] C. Nwankpa, W. Ijomah, A. Gachagan, and S. Marshall, "Activation Functions: Comparison of trends in Practice and Research for Deep Learning," Nov. 2018, [Online]. Available: http://arxiv.org/abs/1811.03378
[19] P. Yu, V. Y. Cui, and J. Guan, "Text Classification by using Natural Language Processing," in IOP Conference Series: Earth and Environmental Science, Mar. 2021, vol. 1802, no. 4. doi: 10.1088/1742-6596/1802/4/042010.
[20] S. Aman and S. Szpakowicz, "Identifying expressions of emotion in text," in Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2007, vol. 4629 LNAI, pp. 196–205. doi: 10.1007/978-3-540-74628-7_27.
[21] M. Banko and E. Brill, "Mitigating the Paucity-of-Data Problem: Exploring the Effect of Training Corpus Size on Classifier Performance for Natural Language Processing."
[22] B. Li, Y. Hou, and W. Che, "Data augmentation approaches in natural language processing: A survey," AI Open, Mar. 2022, doi: 10.1016/j.aiopen.2022.03.001.
[23] J. Wei and K. Zou, "EDA: Easy Data Augmentation Techniques for Boosting Performance on Text Classification Tasks," Jan. 2019, [Online]. Available: http://arxiv.org/abs/1901.11196
[24] A. Karimi, L. Rossi, and A. Prati, "AEDA: An Easier Data Augmentation Technique for Text Classification," Aug. 2021, [Online]. Available: http://arxiv.org/abs/2108.13230
[25] D. R. Beddiar, M. S. Jahan, and M. Oussalah, "Data expansion using back translation and paraphrasing for hate speech detection," Online Social Networks and Media, vol. 24, Jul. 2021, doi: 10.1016/j.osnem.2021.100153.
[26] A. Sherstinsky, "Fundamentals of Recurrent Neural Network (RNN) and Long Short-Term Memory (LSTM) Network," Aug. 2018, doi: 10.1016/j.physd.2019.132306.
[27] S. Hochreiter and J. Schmidhuber, “Long Short-Term Memory,” Neural Computation, vol. 9, no. 8, pp. 1735–1780, Nov. 1997, doi: 10.1162/neco.1997.9.8.1735.
[28] A. Adhikari, A. Ram, R. Tang, and J. Lin, "DocBERT: BERT for Document Classification," Apr. 2019, [Online]. Available: http://arxiv.org/abs/1904.08398
[29] J. S. Lee and J. Hsiang, "PatentBERT: Patent Classification with Fine-Tuning a pre-trained BERT Model," 2019, Accessed: May 21, 2022. [Online]. Available: https://arxiv.org/abs/1906.02124
[30] B. Lee and Y. S. Choi, "Graph Based Network with Contextualized Representations of Turns in Dialogue," Sep. 2021, [Online]. Available: http://arxiv.org/abs/2109.04008

QR CODE