簡易檢索 / 詳目顯示

研究生: 周伯冠
JHOU, BO-GUAN
論文名稱: 使用深度學習方法於產品評論之建議探勘
Mining Suggestions from Product Reviews by Deep Learning Methods
指導教授: 侯文娟
Hou, Wen-Juan
學位類別: 碩士
Master
系所名稱: 資訊工程學系
Department of Computer Science and Information Engineering
論文出版年: 2019
畢業學年度: 107
語文別: 中文
論文頁數: 71
中文關鍵詞: 深度學習產品評論建議探勘詞向量詞頻率
英文關鍵詞: Deep Learning, Product Reviews, Suggestion Mining, Word Vector, Word Frequency
DOI URL: http://doi.org/10.6345/NTNU201900799
論文種類: 學術論文
相關次數: 點閱:155下載:23
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 隨著網路的普及,許多的資訊愈來愈容易被搜索。其中,產品評論相關的訊息也隨著社群網路的發展被公開在社群中。這類的社群包含,論壇、社群網站、和產品官方網站等,甚至有組織專門在蒐集這些評論後組成評論相關的網站,並且將評論分類,給受用戶查看。消費者可以在上述的網站中查看產品的使用狀況與心得和是否符合自身所需,再決定是否購買;產品提供者也可以透過評論持續收集使用者的使用狀況與心得,對產品進行迭代設計,對產品進行改良以符合大眾的需求。

    本研究將評論分為建議句 (suggestion)和非建議句 (non-suggestion)。使用 Stanford Core NLP 斷詞系統將文本以詞為單位進行處理;詞的表達方式分為兩種,詞向量與詞頻率;模型採用深度學習類神經網路,分為全連結類神經網路 (Full Connected Neural Networks, FCNN)、卷積類神經網路 (Convolutional Neural Networks, CNN)與長短期記憶類神經網路 (Long Short-Term Memory, LSTM)。本研究以 Z-Score 方式標準化詞頻率表達式,並用全連結類神經網路訓練,此架構即可和利用詞向量在卷積類神經網路與長短期記憶類神經網路有差不多效果,但速度上快非常多。此外,本研究提供結合詞向量與詞頻率表達式在上述三種模型上訓練,進而對結果做比較分析。效能的評估方式以精準率 (Precision)、召回率 (Recall) 和 F1 分數 (F1-measure, F1)作比較。

    With the popularity of the Internet, much of the information are searched easily. The information related to product reviews is also pushed to the social community, such as forums, social media, and official product websites. There are some organizations that collect the comments to form a review-based website and classify the comments to potential customers. Customers can see the usage and experience of the products to determine whether the products satisfy their needs to decide whether they purchase the products. The product providers can collect the user's usage and experience through comments, and iterate the product design to improve the product and satisfy the needs of customers.

    This study divided comments into two class, suggestion and non-suggestion. The Stanford Core NLP can segment corpus in word units. The expression of words were divided into two types, word vector and word frequency. The model uses deep learning neural network, which include Fully Connected Neural Networks (FCNN), Convolutional Neural Networks (CNN) and Long-Short Term Memory (LSTM) Neural Networks. This study normalized the word frequency in the Z-Score and trained it with a fully connected neural network. This architecture's performance can be similar to the architecture of word vectors in convolutional neural networks and long-short term memory neural networks, but the speed is very fast. In addition, this study provides a combination of word vector and word frequency. The method of performance evaluation is based on Precision, Recall, and F1-measure (F1).

    摘要 I 目錄 V 附表目錄 VII 附圖目錄 VIII 第一章 緒論 1 第一節 研究背景 1 第二節 研究目的 2 第三節 研究意義 2 第四節 論文架構 3 第二章 文獻探討 4 第一節 問題定義 4 第二節 規則式 5 第三節 因子分解機 6 第四節 機器學習 6 第五節 類神經網路 7 第三章 研究步驟與方法 9 第一節 研究步驟 9 第二節 資料收集 15 第三節 前處理 - 標記 16 第四節 前處理 - 斷句 19 第五節 前處理 - 斷詞 19 第六節 詞表達 - 詞向量 20 第七節 詞表達 - 詞頻率 23 第八節 詞表達 – 詞向量結合詞頻率 25 第九節 類神經網路模型 - 全連結類神經網路 26 第十節 類神經網路模型 - 卷積類神經網路 29 第十一節 類神經網路模型 - 長短期記憶類神經網路 32 第十二節 類神經網路架構 36 第十三節 評估方式 42 第四章 實驗結果與討論 44 第一節 實驗資料 44 第二節 實驗結果 47 第三節 分析與討論 61 第五章 結論與未來展望 67 參考文獻 69

    Brun, C., & Hagege, C. (2013). Suggestion Mining: Detecting Suggestions for Improvement in Users' Comments. Research in Computing Science, 70(79.7179), 5379-62.
    Bengio, Y., Ducharme, R., Vincent, P., & Jauvin, C. (2003). A neural probabilistic language model. Journal of machine learning research, 3(Feb), 1137-1155.
    Dong, L., Wei, F., Duan, Y., Liu, X., Zhou, M., & Xu, K. (2013, June). The automated acquisition of suggestions from tweets. In Twenty-Seventh AAAI Conference on Artificial Intelligence.
    Fernández, A. M., Esuli, A., & Sebastiani, F. (2016). Distributional Correspondence Indexing for Cross-Lingual and Cross-Domain Sentiment Classification. Journal of
    artificial intelligence research, 55, 131-163.
    Golchha, H., Gupta, D., Ekbal, A., & Bhattacharyya, P. (2018). Helping each Other: A Framework for Customer-to-Customer Suggestion Mining using a Semi-supervised Deep Neural Network. arXiv preprint arXiv:1811.00379.
    Kim, Y. (2014). Convolutional neural networks for sentence classification. arXiv preprint arXiv:1408.5882.
    Liu, B. (2012). Sentiment analysis and opinion mining. Synthesis lectures on human language technologies, 5(1), 1-167.
    Mikolov, T., Sutskever, I., Chen, K., Corrado, G. S., & Dean, J. (2013a). Distributed representations of words and phrases and their compositionality. In Advances in neural information processing systems (pp. 3111-3119).
    Mikolov, Tomas, et al. "Efficient estimation of word representations in vector space." arXiv preprint arXiv:1301.3781 (2013b).
    Negi, S., & Buitelaar, P. (2015). Towards the extraction of customer-to-customer suggestions from reviews. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (pp. 2159-2167).
    Negi, S., Asooja, K., Mehrotra, S., & Buitelaar, P. (2016). A study of suggestions in opinionated texts and their automatic detection. In Proceedings of the Fifth Joint Conference on Lexical and Computational Semantics (pp. 170-178).
    Pei, J., Han, J., Mortazavi-Asl, B., Wang, J., Pinto, H., Chen, Q., ... & Hsu, M. C. (2004). Mining sequential patterns by pattern-growth: The prefixspan approach. IEEE Transactions on knowledge and data engineering, 16(11), 1424-1440.
    Pennington, J., Socher, R., & Manning, C. (2014). Glove: Global vectors for word representation. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP) (pp. 1532-1543).
    Ramanand, J., Bhavsar, K., & Pedanekar, N. (2010, June). Wishful thinking: finding suggestions and'buy'wishes from product reviews. In Proceedings of the NAACL HLT 2010 workshop on computational approaches to analysis and generation of emotion in text (pp. 54-61). Association for Computational Linguistics.
    Rendle, Steffen. "Factorization machines with libfm." ACM Transactions on Intelligent
    Systems and Technology (TIST) 3.3 (2012): 57.

    下載圖示
    QR CODE