研究生: |
陳煒鈞 Chen, Wei-Jiun |
---|---|
論文名稱: |
基於集成學習方法進行謠言偵測 Using Ensemble Learning Methods on Social Media Rumours Detection |
指導教授: |
侯文娟
Hou, Wen-Juan |
口試委員: |
方瓊瑤
Fang, Chiung-Yao 郭俊桔 Kuo, June-Jei 侯文娟 Hou, Wen-Juan |
口試日期: | 2022/06/20 |
學位類別: |
碩士 Master |
系所名稱: |
資訊工程學系 Department of Computer Science and Information Engineering |
論文出版年: | 2022 |
畢業學年度: | 110 |
語文別: | 中文 |
論文頁數: | 44 |
中文關鍵詞: | 語言模型 、深度學習 、假新聞 、集成學習 |
英文關鍵詞: | Language Model, Deep Learning, Fake news, Ensemble Learning |
研究方法: | 實驗設計法 |
DOI URL: | http://doi.org/10.6345/NTNU202200772 |
論文種類: | 學術論文 |
相關次數: | 點閱:95 下載:0 |
分享至: |
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
網路社交媒體充斥著假消息,連牛津辭典在2016年都將"Post-Truth"列為一個詞彙,錯誤的資訊可能對人造成危害,所以建構一個能夠辨識網路上各種不一樣說法、消息的系統是一個重要的議題。本研究利用預訓練語言模型搭配文字以外的特徵建立出一套辨識謠言的系統,辨識在社交媒體Twitter及Reddit使用者發表內容的真偽。
本論文的資料集來自SemEval 2019 RumourEval: Determining rumour veracity and support for rumours (SemEval 2019 Task 7)的任務B,該任務將Twitter及Reddit上的句子經由人工標註分為3類,真(True)、假(False)、未驗證(Unverified),本研究先經由資料增強的方式增加資料量,接著以不同的語言模型(RoBERTa、ALBERT)及傳統分類(SVM)個別進行訓練,再將不同的模型組合進行集成學習(Ensemble Learning),訓練並給予不同的權重,最後加上後處理達到Marco F1 72 %,RMSE 0.5879的成績。
Media is full of false claims. Even Oxford Dictionaries named “post-truth” as the word in 2016. Misinformation can be harmful to people, so constructing a system that can identify true/false news and statements is an important issue. In this thesis, we use a pretraining language model and some external features to build a system that can identify rumours that are published by users on social media Twitter and Reddit.
In this study, the dataset for our experiments is from task B of SemEval 2019 RumourEval: Determining rumour veracity and support for rumours (SemEval 2019 Task 7). The task divides sentences on Twitter and Reddit into three categories by human annotation, and the label of sentences can be "True","False",and "Unverified". Our research first increases the amount of data by means of data augmentation. Secondly, we train different language models (RoBERTa、ALBERT) and traditional classifiers (SVM) individually. Next, different models are combined for ensemble learning and different weights are given through training. After applying post-processing, the scores of Macro F1 72% and RMSE 0.5879 are achieved.
[1] C.Castillo, M.Mendoza, and B.Poblete. Information credibility on twitter. In
Proceedings of the 20th international conference on World wide web, pages
675–684, 2011, ACM.
[2] Genevieve Gorrell, Elena Kochkina, Maria Liakata, Ahmet Aker, Arkaitz
Zubiaga, Kalina Bontcheva, Leon Derczynski.SemEval-2019 Task 7: RumourEval, Determining Rumour Veracity and Support for Rumours. In
Proceedings of the 13th International Workshop on Semantic Evaluation,Pages845–854,Minneapolis, Minnesota, USA,June 2019. Association for
Computational Linguistics
[3] D. Opitz ,R. Maclin.Popular Ensemble Methods: An Empirical Study. In Journal of Artificial Intelligence Research 11 pp. 169-198. Aug 1, 1999
[4] Leon Derczynski, Kalina Bontcheva, Maria Liakata, Rob Procter, Geraldine
Wong Sak Hoi, Arkaitz Zubiaga. SemEval-2017 Task 8: RumourEval: Determining rumour veracity and support for rumours. In Proceedings of the
11th International Workshop on Semantic Evaluation (SemEval-2017), Pages
69–76.Vancouver, Canada.August 2017. Association for Computational Linguistics
[5] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones,
Aidan N. Gomez, Lukasz Kaiser, Illia Polosukhin. Attention is All You Need. In
Advances in Neural Information Processing Systems, pages 5998-6008. Curran
Associates, Inc. 2017
[6] Elman, Jeffrey L. Finding structure in time In Cognitive science,pages 179-211,
1990. Wiley Online Library
[7] Ilya Sutskever and Oriol Vinyals and Quoc V. Le. Sequence to Sequence Learning with Neural Networks. 2014. arXiv:1409.3215
[8] Mikolov, Tomas and Chen, Kai and Corrado, Greg and Dean, Jeffrey. Efficient estimation of word representations in vector space. In arXiv preprint
arXiv:1301.3781 2013
[9] M. Peters, M. Neumann, M. Iyyer, M. Gardner, C. Clark, K. Lee, and L.
Zettlemoyer, Deep contextualized word representations. In Proceedings of the
2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 2227–2237, 2018.
[10] J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, BERT: Pre-training of
deep bidirectional transformers for language understanding. In Proceedings of
the 2019 Conference of the North American Chapter of the Association for
Computational Linguistics: Human Language Technologies, 2019
[11] S. R. Bowman, G. Angeli, C. Potts, and C. D. Manning, A large annotated
corpus for learning natural language inference. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, (Lisbon, Portugal), pages 632– 642,2015. Association for Computational Linguistics, Sept
[12] A. Radford, K. Narasimhan, T. Salimans, and I. Sutskever, Improving language understanding by generative pre-training, 2018.
[13] M. Fajcik, L. Burget and P. Smrz. BUT-FIT at SemEval-2019 Task 7: Determining the Rumour Stance with Pre-Trained Deep Bidirectional Transformers.
In Proceedings of the 13th International Workshop on Semantic Evaluation 13
(2019), pages 1097-1104,2019.
[14] ] M. Peters, M. Neumann, M. Iyyer, M. Gardner, C. Clark, K. Lee,
and L. Zettlemoyer. Deep contextualized word representations. preprint
arXiv:1802.05365, 2018.
[15] A. Radford, K. Narasimhan, T. Salimans, and I. Sutskever. Improving language understanding with unsupervised learning. OpenAI, 2018.
[16] Q. Li, Q. Zhang and L. Si. eventAI at SemEval-2019 Task 7: Rumor Detection
on Social Media by Exploiting Content, User Credibility and Propagation
Information. In Proceedings of the 13th International Workshop on Semantic
Evaluation, pages 855–859, 2019.
[17] J. A. Meaney, Steven Wilson, Luis Chiruzzo, Adam Lopez, and Walid Magdy.
2021. SemEval 2021 Task 7: HaHackathon, Detecting and Rating Humor and
Offense. In Proceedings of the 15th International Workshop on Semantic Evaluation (SemEval-2021), pages 105–119, Online. Association for Computational
Linguistics.
[18] Bingyan Song, Chunguang Pan, Shengguang Wang, and Zhipeng Luo. 2021.
DeepBlueAI at SemEval-2021 Task 7: Detecting and Rating Humor and Offense with Stacking Diverse Language Model-Based Methods. In Proceedings
of the 15th International Workshop on Semantic Evaluation (SemEval-2021),
pages 1130–1134, Online. Association for Computational Linguistics.
[19] Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi
Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, Veselin Stoyanov.
RoBERTa: A Robustly Optimized BERT Pretraining Approach. July 2019.
arXiv:1907.11692
[20] Lan, Zhenzhong, Chen Mingda, Goodman Sebastian, Gimpel Kevin, Sharma
Piyush, Soricut Radu. ALBERT: A Lite BERT for Self-supervised Learning
of Language Representations. September 2019. arXiv:1909.11942