請用此 Handle URI 來引用此文件:
http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/90792
完整後設資料紀錄
DC 欄位 | 值 | 語言 |
---|---|---|
dc.contributor.advisor | 李育杰 | zh_TW |
dc.contributor.advisor | Yuh-Jye Lee | en |
dc.contributor.author | 李品澤 | zh_TW |
dc.contributor.author | Pin-Zu Li | en |
dc.date.accessioned | 2023-10-03T17:38:32Z | - |
dc.date.available | 2023-11-09 | - |
dc.date.copyright | 2023-10-03 | - |
dc.date.issued | 2023 | - |
dc.date.submitted | 2023-08-08 | - |
dc.identifier.citation | [1] L. E. Boehm. The validity effect: A search for mediating variables. Personality and Social Psychology Bulletin, 20(3):285–293, 1994.
[2] S. Chopra. Towards automatic identification of fake news: Headline-article stance detection with lstm attention models. 2017. [3] J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota, June 2019. Association for Computa-tional Linguistics. [4] Y. Ding, B. Guo, Y. Liu, Y. Liang, H. Shen, and Z. Yu. Metadetector: Meta event knowledge transfer for fake news detection. ACM Trans. Intell. Syst. Technol., 13(6), sep 2022. [5] Y. Dun, K. Tu, C. Chen, C. Hou, and X. Yuan. Kan: Knowledge-aware attention network for fake news detection. Proceedings of the AAAI Conference on Artificial Intelligence, 35(1):81–89, May 2021. [6] N. Ebadi, M. Jozani, K.-K. R. Choo, and P. Rad. A memory network information retrieval model for identification of news misinformation. IEEE Transactions on Big Data, 8(5):1358–1370, 2022. [7] Y. Ganin, E. Ustinova, H. Ajakan, P. Germain, H. Larochelle, F. Laviolette, M. Marc-hand, and V. Lempitsky. Domain-adversarial training of neural networks, 2016. [8] C. Hansen, C. Hansen, and L. C. Lima. Automatic fake news detection: Are mod-els learning to reason? In Annual Meeting of the Association for Computational Linguistics, 2021. [9] K. He, H. Fan, Y. Wu, S. Xie, and R. Girshick. Momentum contrast for unsupervised visual representation learning, 2020. [10] S. Hochreiter and J. Schmidhuber. Long short-term memory. Neural Comput., 9(8):1735–1780, nov 1997. [11] G. Izacard, M. Caron, L. Hosseini, S. Riedel, P. Bojanowski, A. Joulin, and E. Grave. Unsupervised dense information retrieval with contrastive learning, 2021. [12] F. Jin, J. Lu, J. Zhang, and C. Zong. Instance-aware prompt learning for language understanding and generation, 2022. [13] Y. Kim. Convolutional neural networks for sentence classification, 2014. [14] B. Lester, R. Al-Rfou, and N. Constant. The power of scale for parameter-efficient prompt tuning. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 3045–3059, Online and Punta Cana, Dominican Republic, Nov. 2021. Association for Computational Linguistics. [15] X. L. Li and P. Liang. Prefix-tuning: Optimizing continuous prompts for gen-eration. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 4582–4597, Online, Aug. 2021. Association for Computational Linguistics. [16] H. Lin, J. Ma, L. Chen, Z. Yang, M. Cheng, and G. Chen. Detect rumors in microblog posts for low-resource domains via adversarial contrastive learning, 2022. [17] H. Lin, J. Ma, M. Cheng, Z. Yang, L. Chen, and G. Chen. Rumor detection on Twitter with claim-guided hierarchical graph attention networks. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 10035–10047, Online and Punta Cana, Dominican Republic, Nov. 2021. Association for Computational Linguistics. [18] H. Liu, W. Wang, Y. Wang, H. Liu, Z. Liu, and J. Tang. Mitigating gender bias for neural dialogue generation with adversarial learning, 2020. [19] X. Liu, K. Ji, Y. Fu, W. L. Tam, Z. Du, Z. Yang, and J. Tang. P-tuning v2: Prompt tuning can be comparable to fine-tuning universally across scales and tasks, 2022. [20] X. Liu, T. Sun, X. Huang, and X. Qiu. Late prompt tuning: A late prompt could be better than many prompts, 2022. [21] X. Liu, Y. Zheng, Z. Du, M. Ding, Y. Qian, Z. Yang, and J. Tang. Gpt understands, too, 2021. [22] I. Loshchilov and F. Hutter. Decoupled weight decay regularization, 2019. [23] Y.-J. Lu and C.-T. Li. GCAN: Graph-aware co-attention networks for explainable fake news detection on social media. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 505–514, Online, July 2020. Association for Computational Linguistics. [24] G. Ma, C. Hu, L. Ge, and H. Zhang. Open-topic false information detection on social networks with contrastive adversarial learning. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 2911–2923, Abu Dhabi, United Arab Emirates, Dec. 2022. Association for Computational Linguistics. [25] G. Ma, C. Hu, L. Ge, and H. Zhang. Open-topic false information detection on social networks with contrastive adversarial learning. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 2911–2923, Abu Dhabi, United Arab Emirates, Dec. 2022. Association for Computational Linguistics. [26] J. Ma, W. Gao, S. Joty, and K.-F. Wong. Sentence-level evidence embedding for claim verification with hierarchical attention networks. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 2561–2571, Florence, Italy, July 2019. Association for Computational Linguistics. [27] C. J. Maddison, A. Mnih, and Y. W. Teh. The concrete distribution: A continuous relaxation of discrete random variables, 2017. [28] M. P. Mithun, S. Suntwal, and M. Surdeanu. Students who study together learn better: On the importance of collective knowledge distillation for domain transfer in fact verification. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 6968–6973, Online and Punta Cana, Dominican Republic, Nov. 2021. Association for Computational Linguistics. [29] Q. Nan, D. Wang, Y. Zhu, Q. Sheng, Y. Shi, J. Cao, and J. Li. Improving fake news detection of influential domain via domain- and instance-level transfer. In Proceedings of the 29th International Conference on Computational Linguistics, pages 2834–2848, Gyeongju, Republic of Korea, Oct. 2022. International Commit-tee on Computational Linguistics. [30] OpenAI. Gpt-4 technical report, 2023. [31] K. Pelrine, J. Danovitch, and R. Rabbany. The surprising performance of simple baselines for misinformation detection, 2021. [32] K. Popat, S. Mukherjee, A. Yates, and G. Weikum. DeClarE: Debunking fake news and false claims using evidence-aware deep learning. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 22–32, Brussels, Belgium, Oct.-Nov. 2018. Association for Computational Linguistics. [33] V. Pérez-Rosas, B. Kleinberg, A. Lefevre, and R. Mihalcea. Automatic detection of fake news, 2017. [34] P. Qi, J. Cao, X. Li, H. Liu, Q. Sheng, X. Mi, Q. He, Y. Lv, C. Guo, and Y. Yu. Improving fake news detection by using an entity-enhanced framework to fuse di-verse multimodal clues. In Proceedings of the 29th ACM International Conference on Multimedia, MM ’21, page 1212–1220, New York, NY, USA, 2021. Association for Computing Machinery. [35] P. Qi, J. Cao, T. Yang, J. Guo, and J. Li. Exploiting multi-domain visual information for fake news detection. In 2019 IEEE International Conference on Data Mining (ICDM), pages 518–527, 2019. [36] L. Shang, Y. Zhang, Z. Yue, Y. Choi, H. Zeng, and D. Wang. A knowledge-driven domain adaptive approach to early misinformation detection in an emergent health domain on social media. 2022 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM), pages 34–41, 2022. [37] Q. Sheng, J. Cao, X. Zhang, R. Li, D. Wang, and Y. Zhu. Zoom out and observe: News environment perception for fake news detection. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 4543–4556, Dublin, Ireland, May 2022. Association for Computa-tional Linguistics. [38] Q. Sheng, J. Cao, X. Zhang, X. Li, and L. Zhong. Article reranking by memory-enhanced key sentence matching for detecting previously fact-checked claims. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 5468–5481, Online, Aug. 2021. As-sociation for Computational Linguistics. [39] Q. Sheng, X. Zhang, J. Cao, and L. Zhong. Integrating pattern- and fact-based fake news detection via model preference learning. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management. ACM, oct 2021 [40] A. Silva, L. Luo, S. Karunasekera, and C. Leckie. Embracing domain differences in fake news: Cross-domain fake news detection using multi-modal data. Proceedings of the AAAI Conference on Artificial Intelligence, 35:557–565, 05 2021. [41] A. Soleimani, C. Monz, and M. Worring. Bert for evidence retrieval and claim verification. In J. M. Jose, E. Yilmaz, J. Magalhães, P. Castells, N. Ferro, M. J. Silva, and F. Martins, editors, Advances in Information Retrieval, pa[42] J. Thorne, A. Vlachos, C. Christodoulopoulos, and A. Mittal. FEVER: a large-scale dataset for fact extraction and VERification. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 809–819, New Or-leans, Louisiana, June 2018. Association for Computational Linguistics. [43] A. van den Oord, Y. Li, and O. Vinyals. Representation learning with contrastive predictive coding, 2019. [44] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin. Attention is all you need, 2017. [45] N. Vo and K. Lee. Hierarchical multi-head attentive network for evidence-aware fake news detection. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pages 965–975, Online, Apr. 2021. Association for Computational Linguistics. [46] Y. Wang, F. Ma, Z. Jin, Y. Yuan, G. Xun, K. Jha, L. Su, and J. Gao. Eann: Event adversarial neural networks for multi-modal fake news detection. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery Data Mining, KDD’18, page 849–857, New York, NY, USA, 2018. Association for Com-puting Machinery. [47] Y. Wang, F. Ma, H. Wang, K. Jha, and J. Gao. Multimodal emergent fake news detection via meta neural process networks. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining. ACM, aug 2021. [48] Z. Wu, S. Wang, J. Gu, R. Hou, Y. Dong, V. Vydiswaran, and H. Ma. IDPG: An instance-dependent prompt generation method. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 5507–5521, Seattle, United States, July 2022. Association for Computational Linguistics. [49] Z. Yang, J. Ma, H. Chen, H. Lin, Z. Luo, and Y. Chang. A coarse-to-fine cas-caded evidence-distillation neural network for explainable fake news detection. In Proceedings of the 29th International Conference on Computational Linguistics, pages 2608–2621, Gyeongju, Republic of Korea, Oct. 2022. International Commit-tee on Computational Linguistics. [50] Q. Ying, X. Hu, Y. Zhou, Z. Qian, D. Zeng, and S. Ge. Bootstrapping multi-view representations for fake news detection, 2022. [51] C. Yuan, Q. Ma, W. Zhou, J. Han, and S. Hu. Early detection of fake news by utilizing the credibility of news, publishers, and users based on weakly supervised learning. In Proceedings of the 28th International Conference on Computational Linguistics, pages 5444–5454, Barcelona, Spain (Online), Dec. 2020. International Committee on Computational Linguistics. [52] X. Zhang, J. Cao, X. Li, Q. Sheng, L. Zhong, and K. Shu. Mining dual emotion for fake news detection. In Proceedings of the Web Conference 2021. ACM, apr 2021. [53] X. Zhou, A. Jain, V. V. Phoha, and R. Zafarani. Fake news early detection: A theory-driven model. 1(2), jun 2020. [54] X. Zhou and R. Zafarani. A survey of fake news: Fundamental theories, detection methods, and opportunities. ACM Comput. Surv., 53(5), sep 2020. [55] Y. Zhu, Q. Sheng, J. Cao, S. Li, D. Wang, and F. Zhuang. Generalizing to the future. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM, jul 2022. | - |
dc.identifier.uri | http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/90792 | - |
dc.description.abstract | 由於假新聞的多樣性和不斷變化的主題,在早期偵測假新聞面臨重大挑戰。現有方法依賴於難以獲取的特徵或排除特定領域的特徵,這可能會限制其性能。在本文中,我們介紹了一種新穎的方法,利用相關新聞作為參考來早期識別假新聞。我們的方法利用軟提示調整生成兩種特徵:跨不同領域捕捉假新聞共同特徵的領域不變特徵,以及透過假新聞相關文章產生的參考特徵。接著,我們通過動態調整兩種特徵的比例來生成假新聞特徵,並以此來判別假新聞。我們的方法可以以零樣本的方式適應不同的主題或時期,而無需人工製作或難以獲取的特徵。為了評估我們方法的有效性,我們在包含中文和英文數據的兩個數據集上進行了實驗。結果表明,我們的方法在虛假新聞早期檢測方面超過了最先進的方法。 | zh_TW |
dc.description.abstract | Detecting fake news in its early stage poses a significant challenge due to its diverse nature and ever-changing topics. Existing methods rely on difficult-to-acquire features or eliminate domain-specific features, which may limit their performance. In this paper, we introduce a novel method that utilizes related news as references to identify fake news at an early stage. Our approach leverages soft prompt tuning to generate two features: domain-invariant features that capture common characteristics of fake news across various domains and reference features that capture the specific context of each fake news instance with external reference articles. Next, we generate fake news features by dynamically adjusting proportions of the two types of features and use them to determine fake news. Our method can adapt to different topics or periods in a zero-shot manner without needing hand-crafted or hard-to-get features. To evaluate the effectiveness of our approach, we conduct experiments on two datasets comprising Chinese and English language data. The results demonstrate that our method surpasses state-of-the-art techniques in fake news early detection. | en |
dc.description.provenance | Submitted by admin ntu (admin@lib.ntu.edu.tw) on 2023-10-03T17:38:32Z No. of bitstreams: 0 | en |
dc.description.provenance | Made available in DSpace on 2023-10-03T17:38:32Z (GMT). No. of bitstreams: 0 | en |
dc.description.tableofcontents | Acknowledgements i
摘要 iii Abstract v Contents vii List of Figures xi List of Tables xiii Denotation xv Chapter 1 Introduction 1 Chapter 2 Related Work 5 2.1 Content-based Fake News Detection . . . . . . . . . . . . . . . . . . 5 2.1.1 Input Data Perspective . . . . . . . . . . . . . . . . . . . . . . . . 5 2.1.2 Application Scenario . . . . . . . . . . . . . . . . . . . . . . . . . 6 2.2 Prompt Tuning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 Chapter 3 Methodology 9 3.1 Problem Formulation . . . . . . . . . . . . . . . . . . . . . . . . . . 9 3.2 Proposed Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 3.2.1 Dual-tower Dense Retriever - Contriever . . . . . . . . . . . . . . . 11 3.2.2 Domain-Invariant Encoder . . . . . . . . . . . . . . . . . . . . . . 12 3.2.3 Knowledge-aware Encoder . . . . . . . . . . . . . . . . . . . . . . 14 3.2.4 Shared-domain Classifier . . . . . . . . . . . . . . . . . . . . . . . 16 3.2.5 CNN Token-level Fusion Gate . . . . . . . . . . . . . . . . . . . . 18 3.2.6 Detector Module . . . . . . . . . . . . . . . . . . . . . . . . . . . 20 3.2.7 Training Objective and Inference Stage . . . . . . . . . . . . . . . . 20 Chapter 4 Experiments 21 4.1 Dataset . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21 4.1.1 Cofacts Dataset . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21 4.1.1.1 Cofacts Reference Articles . . . . . . . . . . . . . . . 22 4.1.2 NEP-eng Dataset . . . . . . . . . . . . . . . . . . . . . . . . . . . 22 4.1.2.1 NEP-eng Reference Articles . . . . . . . . . . . . . . . 22 4.2 Experimental Setup . . . . . . . . . . . . . . . . . . . . . . . . . . . 23 4.2.1 Evaluation Metrics . . . . . . . . . . . . . . . . . . . . . . . . . . 23 4.2.2 Baseline Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23 4.2.2.1 Pretrained Models . . . . . . . . . . . . . . . . . . . . 24 4.2.2.2 Cross-domain Methods . . . . . . . . . . . . . . . . . 24 4.2.2.3 Reference-based Methods . . . . . . . . . . . . . . . . 24 4.2.3 Large Language Models . . . . . . . . . . . . . . . . . . . . . . . 25 4.2.4 Implementation Details . . . . . . . . . . . . . . . . . . . . . . . . 26 4.2.5 Dense Retreiver . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26 4.2.6 DR-FEND . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26 4.2.6.1 Hyperparameters Search . . . . . . . . . . . . . . . . . 28 4.3 Performance Comparison . . . . . . . . . . . . . . . . . . . . . . . . 29 4.3.1 Ablation Study . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30 4.3.1.1 Effectiveness of Using MHA Late Prompt . . . . . . . 31 4.3.1.2 Effectiveness of Using Knowledge-aware Prompt . . . 31 4.3.1.3 Effectiveness of Shared-domain Classifier . . . . . . . 33 4.3.1.4 Effectiveness of CNN Token-level Fusion Gate . . . . 34 4.3.2 Case Study . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 4.3.2.1 Case A . . . . . . . . . . . . . . . . . . . . . . . . . . 35 4.3.2.2 Case B . . . . . . . . . . . . . . . . . . . . . . . . . . 36 4.3.2.3 Case C . . . . . . . . . . . . . . . . . . . . . . . . . . 36 4.3.2.4 Case D . . . . . . . . . . . . . . . . . . . . . . . . . . 36 Chapter 5 Conclusion 37 References 39 Appendix A — Implementation Details 49 A.1 Cofacts Dataset . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49 A.2 NEP-eng Dataset . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51 Appendix B — Experiments 53 B.1 Baseline Result Using bert-base-chinese on Cofacts . . . . . . . . . . 53 | - |
dc.language.iso | en | - |
dc.title | 利用參考引導和領域不變的後置軟提示進行跨領域假新聞檢測 | zh_TW |
dc.title | Reference-Guided and Domain-Invariant Late Prompts for Fake News Early Detection | en |
dc.type | Thesis | - |
dc.date.schoolyear | 111-2 | - |
dc.description.degree | 碩士 | - |
dc.contributor.coadvisor | 李宏毅 | zh_TW |
dc.contributor.coadvisor | Hung-Yi Lee | en |
dc.contributor.oralexamcommittee | 鮑興國;許永真 | zh_TW |
dc.contributor.oralexamcommittee | Hsing-Kuo Pao;Jane Yung-jen Hsu | en |
dc.subject.keyword | 跨領域假新聞偵測,假新聞早期偵測,提示學習, | zh_TW |
dc.subject.keyword | cross domain fake news detection,fake news early detection,prompt tuning, | en |
dc.relation.page | 53 | - |
dc.identifier.doi | 10.6342/NTU202302479 | - |
dc.rights.note | 同意授權(限校園內公開) | - |
dc.date.accepted | 2023-08-09 | - |
dc.contributor.author-college | 電機資訊學院 | - |
dc.contributor.author-dept | 資料科學學位學程 | - |
顯示於系所單位: | 資料科學學位學程 |
文件中的檔案:
檔案 | 大小 | 格式 | |
---|---|---|---|
ntu-111-2.pdf 授權僅限NTU校內IP使用(校園外請利用VPN校外連線服務) | 2.66 MB | Adobe PDF | 檢視/開啟 |
系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。