簡易檢索 / 詳目顯示

研究生: 陳牧凡
Mu-Fan Chen
論文名稱: 透過像素注意力之細微特徵提取與具區別性之多興趣推薦系統
Fine-grained Feature Extraction via Pixel Attention and Distinctive Interest Learning for Multi-Interest Recommendation
指導教授: 戴碧如
Bi-Ru Dai
口試委員: 陳怡伶
Yi-Ling Chen
戴志華
Chih-Hua Tai
沈之涯
Chih-Ya Shen
戴碧如
Bi-Ru Dai
學位類別: 碩士
Master
系所名稱: 電資學院 - 資訊工程系
Department of Computer Science and Information Engineering
論文出版年: 2023
畢業學年度: 111
語文別: 英文
論文頁數: 65
中文關鍵詞: 推薦系統多興趣對比學習專注力機制
外文關鍵詞: Recommender Systems, Multi-interest, Contrastive Learning, Attention Mechanism
相關次數: 點閱:349下載:0
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 在推薦系統領域中,神經網絡的使用越來越普遍,在過去大部分的研究當中通常透過單一表徵來代表使用者的整體偏好,而在一些研究當中注意到使用者偏好可以被解釋為使用者的多方面的興趣,因此多興趣的概念在推薦系統設計中引起了越來越多的關注。通常這些多興趣框架會使用注意力機制來提取多興趣的表徵。然而,透過注意力機制的多興趣表示無法清楚地展示使用者的各方面興趣。另外它們也忽略了隱藏的細微特徵和使用者歷史序列中不同項目之間的關係。因此,我們提出了一種新的方法,命名為Fine-grained Feature Extraction via Pixel Attention and Distinctive Interest Learning (FEPADI)。其利用額外的注意力網絡來捕捉細微特徵。此外,FEPADI還利用了對比學習來增強使用者興趣表徵的差異性。通過增強使用者興趣表徵,我們可以獲得更準確和詳細的使用者偏好,從而實現更有效的推薦。我們也在數個真實世界資料集上進行了實驗,並將我們的方法與最先進的方法進行了比較。實驗結果顯示,與比較對象相比FEPADI取得了更加優異的表現,並且在額外的實驗中也驗證了FEPADI中每個模塊的有效性。


    Recently, the use of neural networks in the recommender system field has been increasing year by year, and a single embedding is generally used to represent the overall preference of a user. Some researchers further noticed that user preference can also be interpreted as multiple aspects of user interests, and the concept of multi-interests attracts more and more attention in the design of recommender systems. Commonly, those multi-interest frameworks use the attention mechanism to extract multi-interest representations. However, the multi-interest representations obtained by those works cannot clearly show distinguishable aspects of user interests. Moreover, they usually ignore the fine-grained feature hidden in the items and the relationship between different items in the user history sequence. Therefore, we propose a novel method called FEPADI, which leverages an additional attention network for capturing the fine-grained feature. Additionally, contrastive learning is utilized in FEPADI to enhance the distinctiveness of the user interest representations. With the enhancement of the user interest representations, we can obtain more accurate and detailed user preferences, leading to more effective recommendations. We conducted experiments on the real-world datasets and compared our performance with that of state-of-the-art sequential, contrastive learning and multi-interest frameworks and several baseline recommender methods. Experimental results demonstrated the significant improvements of FEPADI compared with the competitors and validated the effectiveness of each module in the proposed FEPADI.

    Abstract in Chinese iii Abstract in English iv Acknowledgements v Contents vi List of Figures ix List of Tables x List of Algorithms xi 1 Introduction 1 2 Related Work 4 2.1 Sequential-based Recommendation 4 2.2 Multi-interest Recommendation 5 2.3 Attention Mechanism 6 2.4 Contrastive Learning 7 3 Methodology 8 3.1 Problem Formulation and Model Overview 8 3.2 Multi-Interest Module 11 3.2.1 Embedding Layer 12 3.2.2 Multi-interest Extractor 13 3.2.3 Pixel Attention Module 13 3.3 Contrastive Learning Module 16 3.4 Training and Inference 20 4 Experiments 22 4.1 Experimental Setup 22 4.1.1 Datasets 23 4.1.2 Comparison Methods 25 4.1.3 Parameter Configuration 27 4.1.4 Evaluation Metrics 27 4.2 Overall Performance 28 4.3 Model Analysis 34 4.3.1 Analysis of the number of user interests K 35 4.3.2 Analysis of the contrastive loss coefficient λ 35 4.3.3 Analysis of the positive sample selection threshold σ 36 4.3.4 Analysis of the negative sample selection strategy 37 4.3.5 Ablation Study 40 4.4 Case Study 41 4.4.1 Multi-interest Effectiveness 41 4.4.2 Multi-interest Distinction 44 5 Conclusions 47 References 48 Letter of Authority 53

    [1] B. Hidasi, A. Karatzoglou, L. Baltrunas, and D. Tikk, “Session-based recommendations with recurrent
    neural networks,” arXiv preprint arXiv:1511.06939, 2015.
    [2] P. Covington, J. Adams, and E. Sargin, “Deep neural networks for youtube recommendations,” in
    Proceedings of the 10th ACM conference on recommender systems, pp. 191–198, 2016.
    [3] C. Li, Z. Liu, M. Wu, Y. Xu, H. Zhao, P. Huang, G. Kang, Q. Chen, W. Li, and D. L. Lee, “Multi-
    interest network with dynamic routing for recommendation at Tmall,” in Proceedings of the 28th ACM
    international conference on information and knowledge management, pp. 2615–2623, 2019.
    [4] Y. Cen, J. Zhang, X. Zou, C. Zhou, H. Yang, and J. Tang, “Controllable multi-interest framework for
    recommendation,” in Proceedings of the 26th ACM SIGKDD International Conference on Knowledge
    Discovery & Data Mining, pp. 2942–2951, 2020.
    [5] D. Yin and S. Feng, “Enhanced Attention Framework for Multi-Interest Sequential Recommendation,”
    IEEE Access, vol. 10, pp. 67703–67712, 2022. ISBN: 2169-3536 Publisher: IEEE.
    [6] G. Chen, X. Zhang, Y. Zhao, C. Xue, and J. Xiang, “Exploring periodicity and interactivity in multi-
    interest framework for sequential recommendation,” arXiv preprint arXiv:2106.04415, 2021.
    [7] C. Sun, P. Lu, L. Cheng, Z. Cao, X. Dong, Y. Tang, J. Zhou, and L. Mo, “Multi-interest sequence mod-
    eling for recommendation with causal embedding,” in Proceedings of the 2022 SIAM International
    Conference on Data Mining (SDM), pp. 406–414, SIAM, 2022.
    [8] Z. Wang and Y. Shen, “Time-aware Multi-interest Capsule Network for Sequential Recommendation,”
    in Proceedings of the 2022 SIAM International Conference on Data Mining (SDM), pp. 558–566,
    SIAM, 2022.
    [9] Z. Chai, Z. Chen, C. Li, R. Xiao, H. Li, J. Wu, J. Chen, and H. Tang, “User-Aware Multi-Interest
    Learning for Candidate Matching in Recommenders,” in Proceedings of the 45th International ACM
    SIGIR Conference on Research and Development in Information Retrieval, pp. 1326–1335, 2022.
    [10] Z. Liu, Y. Luo, D. Zeng, Q. Liu, D. Chang, D. Kong, and Z. Chen, “Improving Multi-Interest Network
    with Stable Learning,” arXiv preprint arXiv:2207.07910, 2022.
    [11] S. Zhang, L. Yang, D. Yao, Y. Lu, F. Feng, Z. Zhao, T.-S. Chua, and F. Wu, “Re4: Learning to Re-
    contrast, Re-attend, Re-construct for Multi-interest Recommendation,” in Proceedings of the ACM
    Web Conference 2022, pp. 2216–2226, 2022.
    [12] B. Li, B. Jin, J. Song, Y. Yu, Y. Zheng, and W. Zhuo, “Improving Micro-video Recommendation via
    Contrastive Multiple Interests,” May 2022. arXiv:2205.09593 [cs].
    47
    [13] Y. Wang, Y. Qin, F. Sun, B. Zhang, X. Hou, K. Hu, J. Cheng, J. Lei, and M. Zhang, “DisenCTR:
    Dynamic Graph-based Disentangled Representation for Click-Through Rate Prediction,” in Proceed-
    ings of the 45th International ACM SIGIR Conference on Research and Development in Information
    Retrieval, (Madrid Spain), pp. 2314–2318, ACM, July 2022.
    [14] Y. Tian, J. Chang, Y. Niu, Y. Song, and C. Li, “When Multi-Level Meets Multi-Interest: A Multi-
    Grained Neural Model for Sequential Recommendation,” in Proceedings of the 45th International
    ACM SIGIR Conference on Research and Development in Information Retrieval, (Madrid Spain),
    pp. 1632–1641, ACM, July 2022.
    [15] S. Sabour, N. Frosst, and G. E. Hinton, “Dynamic routing between capsules,” Advances in neural
    information processing systems, vol. 30, 2017.
    [16] R. He and J. McAuley, “Fusing similarity models with markov chains for sparse sequential recom-
    mendation,” in 2016 IEEE 16th international conference on data mining (ICDM), pp. 191–200, IEEE,
    2016.
    [17] S. Rendle, “Factorization machines,” in 2010 IEEE International conference on data mining, pp. 995–
    1000, IEEE, 2010.
    [18] S. Rendle, C. Freudenthaler, and L. Schmidt-Thieme, “Factorizing personalized markov chains for
    next-basket recommendation,” in Proceedings of the 19th international conference on World wide
    web, pp. 811–820, 2010.
    [19] J. Tang and K. Wang, “Personalized top-n sequential recommendation via convolutional sequence
    embedding,” in Proceedings of the eleventh ACM international conference on web search and data
    mining, pp. 565–573, 2018.
    [20] J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for semantic segmentation,”
    in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 3431–3440,
    2015.
    [21] K. Cho, B. Van Merriënboer, C. Gulcehre, D. Bahdanau, F. Bougares, H. Schwenk, and Y. Ben-
    gio, “Learning phrase representations using RNN encoder-decoder for statistical machine translation,”
    arXiv preprint arXiv:1406.1078, 2014.
    [22] K. Zhou, H. Yu, W. X. Zhao, and J.-R. Wen, “Filter-enhanced mlp is all you need for sequential
    recommendation,” in Proceedings of the ACM Web Conference 2022, pp. 2388–2399, 2022.
    [23] Y. Li, T. Chen, P.-F. Zhang, and H. Yin, “Lightweight self-attentive sequential recommendation,” in
    Proceedings of the 30th ACM International Conference on Information & Knowledge Management,
    pp. 967–977, 2021.
    [24] Z. Liu, Z. Fan, Y. Wang, and P. S. Yu, “Augmenting sequential recommendation with pseudo-prior
    items via reversely pre-training transformer,” in Proceedings of the 44th international ACM SIGIR
    conference on Research and development in information retrieval, pp. 1608–1612, 2021.
    48
    [25] F. Sun, J. Liu, J. Wu, C. Pei, X. Lin, W. Ou, and P. Jiang, “Bert4rec: Sequential recommendation with
    bidirectional encoder representations from transformer,” in Proceedings of the 28th ACM international
    conference on information and knowledge management, pp. 1441–1450, 2019.
    [26] W.-C. Kang and J. McAuley, “Self-attentive sequential recommendation,” in 2018 IEEE international
    conference on data mining (ICDM), pp. 197–206, IEEE, 2018.
    [27] S. Liu, Z. Chen, H. Liu, and X. Hu, “User-video co-attention network for personalized micro-video
    recommendation,” in The World Wide Web Conference, pp. 3020–3026, 2019.
    [28] X. Chen, D. Liu, Z.-J. Zha, W. Zhou, Z. Xiong, and Y. Li, “Temporal hierarchical attention at category-
    and item-level for micro-video click-through prediction,” in Proceedings of the 26th ACM interna-
    tional conference on Multimedia, pp. 1146–1153, 2018.
    [29] H. Jiang, W. Wang, Y. Wei, Z. Gao, Y. Wang, and L. Nie, “What aspect do you like: Multi-scale time-
    aware user interest modeling for micro-video recommendation,” in Proceedings of the 28th ACM
    International conference on Multimedia, pp. 3487–3495, 2020.
    [30] T. Wang and P. Isola, “Understanding contrastive representation learning through alignment and uni-
    formity on the hypersphere,” in International Conference on Machine Learning, pp. 9929–9939,
    PMLR, 2020.
    [31] T. Chen, S. Kornblith, M. Norouzi, and G. Hinton, “A Simple Framework for Contrastive Learning
    of Visual Representations,” June 2020. arXiv:2002.05709 [cs, stat].
    [32] K. He, H. Fan, Y. Wu, S. Xie, and R. Girshick, “Momentum contrast for unsupervised visual rep-
    resentation learning,” in Proceedings of the IEEE/CVF conference on computer vision and pattern
    recognition, pp. 9729–9738, 2020.
    [33] Z. Wu, S. Wang, J. Gu, M. Khabsa, F. Sun, and H. Ma, “CLEAR: Contrastive Learning for Sentence
    Representation,” Dec. 2020. arXiv:2012.15466 [cs].
    [34] J. Giorgi, O. Nitski, B. Wang, and G. Bader, “DeCLUTR: Deep Contrastive Learning for Unsupervised
    Textual Representations,” May 2021. arXiv:2006.03659 [cs].
    [35] D. Zhang, F. Nan, X. Wei, S. Li, H. Zhu, K. McKeown, R. Nallapati, A. Arnold, and B. Xiang,
    “Supporting Clustering with Contrastive Learning,” May 2021. arXiv:2103.12953 [cs].
    [36] T. Gao, X. Yao, and D. Chen, “SimCSE: Simple Contrastive Learning of Sentence Embeddings,” May
    2022. arXiv:2104.08821 [cs].
    [37] X. Xie, F. Sun, Z. Liu, S. Wu, J. Gao, J. Zhang, B. Ding, and B. Cui, “Contrastive learning for se-
    quential recommendation,” in 2022 IEEE 38th international conference on data engineering (ICDE),
    pp. 1259–1273, IEEE, 2022.
    [38] S. Zhang, D. Yao, Z. Zhao, T.-S. Chua, and F. Wu, “Causerec: Counterfactual user sequence synthesis
    for sequential recommendation,” in Proceedings of the 44th International ACM SIGIR Conference on
    Research and Development in Information Retrieval, pp. 367–377, 2021.
    49
    [39] Y. Chen, Z. Liu, J. Li, J. McAuley, and C. Xiong, “Intent contrastive learning for sequential recom-
    mendation,” in Proceedings of the ACM Web Conference 2022, pp. 2172–2182, 2022.
    [40] S. Zhang, B. Li, D. Yao, F. Feng, J. Zhu, W. Fan, Z. Zhao, X. He, T.-s. Chua, and F. Wu,
    “CCL4Rec: Contrast over Contrastive Learning for Micro-video Recommendation,” Aug. 2022.
    arXiv:2208.08024 [cs].
    [41] Y. Wei, X. Wang, Q. Li, L. Nie, Y. Li, X. Li, and T.-S. Chua, “Contrastive Learning for Cold-Start
    Recommendation,” in Proceedings of the 29th ACM International Conference on Multimedia, (Virtual
    Event China), pp. 5382–5390, ACM, Oct. 2021.
    [42] H. Zhou, J. Liu, Z. Li, J. Yu, and H. Yang, “Cross-domain User Preference Learning for Cold-start
    Recommendation,” Dec. 2021. arXiv:2112.03667 [cs].
    [43] Y. Yu, B. Jin, J. Song, B. Li, Y. Zheng, and W. Zhuo, “Improving Micro-video Recommendation
    by Controlling Position Bias,” in Machine Learning and Knowledge Discovery in Databases (M.-R.
    Amini, S. Canu, A. Fischer, T. Guns, P. Kralj Novak, and G. Tsoumakas, eds.), vol. 13713, pp. 508–
    523, Cham: Springer International Publishing, 2023. Series Title: Lecture Notes in Computer Science.
    [44] Z. Liu, Y. Ma, Y. Ouyang, and Z. Xiong, “Contrastive Learning for Recommender System,” Jan. 2021.
    arXiv:2101.01317 [cs].
    [45] G. Adomavicius, N. Manouselis, and Y. Kwon, “Multi-criteria recommender systems,” in Recom-
    mender systems handbook, pp. 769–803, Springer, 2010.
    [46] N. Nassar, A. Jafar, and Y. Rahhal, “Multi-criteria collaborative filtering recommender by fusing deep
    neural network and matrix factorization,” Journal of Big Data, vol. 7, no. 1, pp. 1–12, 2020.
    [47] Y. Zhang, Y. Xiong, X. Kong, and Y. Zhu, “Learning node embeddings in interaction graphs,” in
    Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, pp. 397–
    406, 2017.
    [48] H. Zhang, I. Goodfellow, D. Metaxas, and A. Odena, “Self-attention generative adversarial networks,”
    in International conference on machine learning, pp. 7354–7363, PMLR, 2019.
    [49] J. Fu, J. Liu, H. Tian, Y. Li, Y. Bao, Z. Fang, and H. Lu, “Dual attention network for scene seg-
    mentation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition,
    pp. 3146–3154, 2019.
    [50] A. v. d. Oord, Y. Li, and O. Vinyals, “Representation learning with contrastive predictive coding,”
    arXiv preprint arXiv:1807.03748, 2018.
    [51] S. Jean, K. Cho, R. Memisevic, and Y. Bengio, “On using very large target vocabulary for neural
    machine translation,” arXiv preprint arXiv:1412.2007, 2014.
    [52] Y. Bengio and J.-S. Senécal, “Adaptive importance sampling to accelerate training of a neural proba-
    bilistic language model,” IEEE Transactions on Neural Networks, vol. 19, no. 4, pp. 713–722, 2008.
    50
    [53] D. Liang, R. G. Krishnan, M. D. Hoffman, and T. Jebara, “Variational autoencoders for collaborative
    filtering,” in Proceedings of the 2018 world wide web conference, pp. 689–698, 2018.
    [54] J. McAuley, C. Targett, Q. Shi, and A. Van Den Hengel, “Image-based recommendations on styles
    and substitutes,” in Proceedings of the 38th international ACM SIGIR conference on research and
    development in information retrieval, pp. 43–52, 2015.
    [55] M. Wan and J. McAuley, “Item recommendation on monotonic behavior chains,” in Proceedings of the
    12th ACM Conference on Recommender Systems, (Vancouver British Columbia Canada), pp. 86–94,
    ACM, Sept. 2018.
    [56] M. Wan, R. Misra, N. Nakashole, and J. J. McAuley, “Fine-Grained Spoiler Detection from Large-
    Scale Review Corpora,” in Proceedings of the 57th Conference of the Association for Computational
    Linguistics, ACL 2019, Florence, Italy, July 28- August 2, 2019, Volume 1: Long Papers (A. Korho-
    nen, D. R. Traum, and L. Màrquez, eds.), pp. 2605–2610, Association for Computational Linguistics,
    2019.
    [57] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint
    arXiv:1412.6980, 2014.
    [58] T. N. Kipf and M. Welling, “Semi-supervised classification with graph convolutional networks,” arXiv
    preprint arXiv:1609.02907, 2016

    無法下載圖示 全文公開日期 2033/08/27 (校內網路)
    全文公開日期 本全文未授權公開 (校外網路)
    全文公開日期 本全文未授權公開 (國家圖書館:臺灣博碩士論文系統)
    QR CODE