簡易檢索 / 詳目顯示

研究生: 簡子芸
Chien, Tzu-Yun
論文名稱: 基於面向學習之商品評分預測與解釋文本生成模型
A Recommendation Model for Rating Prediction and Explanation Generation via Aspects Learning
指導教授: 柯佳伶
Koh, Jia-Ling
口試委員: 徐嘉連
Hsu, Jia-Lien
吳宜鴻
Wu, Yi-Hong
柯佳伶
Koh, Jia-Ling
口試日期: 2022/01/24
學位類別: 碩士
Master
系所名稱: 資訊工程學系
Department of Computer Science and Information Engineering
論文出版年: 2022
畢業學年度: 110
語文別: 中文
論文頁數: 62
中文關鍵詞: 可解釋性推薦系統多任務學習自然語言生成
英文關鍵詞: Explainable Recommendation system, Multi-task Learning, Natural Language Generation
DOI URL: http://doi.org/10.6345/NTNU202200270
論文種類: 學術論文
相關次數: 點閱:119下載:20
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 本論文提出一個基於面向學習概念的模型,用來進行商品評分預測及對評分的解釋文本生成,稱為LARGE (Learning Aspects-representation for Rating and Generating Explanation)模型。在模型的編碼器中我們設計可學習多面向特徵空間的神經層,由商品評論文內容及使用者嵌入向量學習出對應的面向特徵向量,除了用以提供評分預測,並將商品的面向特徵向量融入每次的解碼狀態,引導生成的評分解釋文本能聚焦於商品所具有的面向。LARGE模型採多任務學習方式進行訓練,透過結合兩個不同目標任務的損失函數進行整體參數優化,且在評分預測的損失函數,加入權重調整策略,以降低推薦系統中評分資料分布不均對預測效能的影響。本論文採用亞馬遜資料集中三個不同商品類別的資料進行測試,實驗結果顯示,LARGE模型比相關研究所提出的代表性模型NRT,有效提升在評分預測及文本解釋生成的效能。此外,LARGE在解釋文本敘述中的類別型面向詞涵蓋率,比需輸入指定面向詞的NRT擴展模型有更高的涵蓋率。

    In this paper, we proposed a recommendation model, called LARGE (Learning Aspects-representation for Rating and Generating Explanation), to perform aspect-based representation learning for both rating prediction and explanation generation. In the encoder, we designed a neural layer to learn multi-aspect representations from item reviews and user embedding for the task of rating prediction. In the task of explanation generation, we fused the learned aspect-based representation of items into each decoding state in order to guide explanation generation focus on the specific aspects of the item. The LARGE model is trained by a multi-task learning approach, where the parameters are tuned by optimizing a linear combination of the loss on the two target tasks. In addition, to reduce the bias of model training due to data unbalance, a weight adjustment strategy is applied to the loss function of rating prediction. The experiments are performed on 3 categories selected from the Amazon review dataset. The result of the experiments shows that the LARGE model significantly outperforms NRT on both tasks. Furthermore, to compare with an extension model of NRT using a given aspect word as model input, the rating explanation generated by LARGE has higher coverage on aspect words.

    第一章 緒論 1 1.1 研究動機與目的 1 1.2 論文方法 3 1.3 論文架構 5 第二章 文獻探討 6 2.1 評分預測之單任務推薦系統 6 2.1.1 基於矩陣分解法的評分預測推薦系統 6 2.1.2 基於深度學習方法的評分預測推薦系統 7 2.2 結合評分解釋生成之多任務推薦系統 11 第三章 問題定義與資料前處理 15 3.1 問題定義 15 3.2 資料前處理 16 第四章 評分預測及評分解釋文本生成模型 18 4.1 評分預測模型 19 4.1.1 嵌入層 19 4.1.2 潛在面向特徵表示法學習 20 4.1.3 潛在面向重要性評估 21 4.1.4 評分預測層 23 4.1.5 損失函數 24 4.2 評分解釋文本生成模型 25 4.2.1 評分解釋生成層 26 4.2.2 損失函數 27 4.3 整體模型之損失函數設計與訓練策略 28 第五章 實驗評估及討論 29 5.1 資料集說明及模型參數設定 29 5.2 評估指標 32 5.3 模型在評分預測任務上的效能評估 38 5.4 模型在推薦解釋文本生成上的效能評估 43 第六章 結論與未來研究方向 53 參考文獻 54 附錄 57

    [1] Yongfeng Zhang, Guokun Lai, Min Zhang, Yi Zhang, Yiqun Liu, and Shaoping Ma. 2014. “Explicit Factor Models for Explainable Recommendation based on Phrase-level Sentiment Analysis,” In SIGIR. ACM, 83–92.
    [2] Chong Chen, Min Zhang, Yiqun Liu, and Shaoping Ma. 2018. “Neural attentional rating regression with review-level explanations,” In Proceedings of the 2018 World Wide Web Conference on World Wide Web. International World Wide Web Conferences Steering Committee, 1583–1592.
    [3] Zhiyong Cheng, Ying Ding, Lei Zhu, and Mohan Kankanhalli. 2018. Aspect-Aware Latent Factor Model: Rating Prediction with Ratings and Reviews. In WWW ’18. 639–648.
    [4] Felipe Costa, Sixun Ouyang, Peter Dolog, and Aonghus Lawlor. 2017. “Automatic Generation of Natural Language Explanations,” In Proceedings of the 23rd International Conference on Intelligent User Interfaces Companion. ACM, 57.
    [5] Li Dong, Shaohan Huang, Furu Wei, Mirella Lapata, Ming Zhou, and Ke Xu. 2017. Learning to Generate Product Reviews from Attributes. In EACL, Vol. 1. 623–632.
    [6] Piji Li, Zihao Wang, Zhaochun Ren, Lidong Bing, and Wai Lam. 2017. Neural rating regression with abstractive tips generation for recommendation. In Proceedings of the 40th International ACM SIGIR conference on Research and Development in Information Retrieval. ACM, 345–354.
    [7] Chen, H., X. Chen, S. Shi, and Y. Zhang. 2019. “Generate natural language explanations for recommendation”. In Proceedings of the SIGIR 2019 Workshop on Explainable Recommendation and Search
    [8] Lei Li, Yongfeng Zhang, and Li Chen. 2020. Generate neural template explanations for recommendation. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management. 755–764.
    [9] Jin Yao Chin, Kaiqi Zhao, Shafiq Joty, and Gao Cong. 2018. ANR: Aspect-based Neural Recommender. In CIKM. 147–156.
    [10] Wang, N., H. Wang, Y. Jia, and Y. Yin . 2018. “Explainable recommendation via multi-task learning in opinionated text data”. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval. ACM.
    [11] Ni, J., J. Li, and J. McAuley (2019). “Justifying recommendations using distantly-labeled reviews and fine-grained aspects”. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). 188–197.
    [12] Yehuda Koren, Robert Bell, and Chris Volinsky. 2009. Matrix Factorization Techniques for Recommender Systems. Computer 42, 8 (Aug. 2009), 30–37.
    [13] Yehuda Koren. 2008. Factorization meets the neighborhood: a multifaceted
    collaborative filtering model. In SIGKDD. 426–434.
    [14] Sungyong Seo, Jing Huang, Hao Yang, and Yan Liu. 2017. Interpretable Convolutional Neural Networks with Dual Local and Global Attention for Review Rating Prediction. In RecSys ’17. ACM, New York, NY, USA, 297–305.
    [15] Lei Zheng, Vahid Noroozi, and Philip S. Yu. 2017. Joint Deep Modeling of Users and Items Using Reviews for Recommendation. In WSDM ’17. ACM, 425–434
    [16] Rose Catherine and William Cohen. 2017. TransNets: Learning to Transform for Recommendation. In RecSys ’17. ACM, New York, NY, USA, 288–296.
    [17] Qiming Diao, Minghui Qiu, Chao-Yuan Wu, Alexander J. Smola, Jing Jiang, and Chong Wang. 2014. Jointly modeling aspects, ratings and sentiments for movie recommendation (JMARS). In KDD. 193–202.
    [18] Yao Wu and Martin Ester. 2015. FLAME: A Probabilistic Model Combining Aspect Based Opinion Mining and Collaborative Filtering. In WSDM ’15. ACM, New York, NY, USA, 199–208.
    [19] Yi Tay, Anh Tuan Luu, and Siu Cheung Hui. 2018. Multi-Pointer Co-Attention Networks for Recommendation. In KDD. 2309–2318.
    [20] Xiangnan He, Lizi Liao, Hanwang Zhang, Liqiang Nie, Xia Hu, and Tat-Seng Chua. 2017. Neural Collaborative Filtering. In Proceedings of the 26th International Conference on World Wide Web (WWW ’17). 173–182.
    [21] Chao-Yuan Wu, Amr Ahmed, Alex Beutel, Alexander J. Smola, and How Jing. 2017. Recurrent Recommender Networks. In Proceedings of the Tenth ACM International Conference on Web Search and Data Mining, WSDM 2017, Cambridge, United Kingdom, February 6-10, 2017. 495–503.
    [22] Nan Wang, Hongning Wang, Yiling Jia, and Yue Yin. 2018. Explainable Recommendation via Multi-Task Learning in Opinionated Text Data. In SIGIR. 165–174.
    [23] Konstantin Bauman, Bing Liu, and Alexander Tuzhilin. 2017. Aspect Based Recommendations: Recommending Items with the Most Valuable Aspects Based on User Reviews. In KDD ’17. ACM, 717–725.

    下載圖示
    QR CODE