請用此 Handle URI 來引用此文件:
http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/97019完整後設資料紀錄
| DC 欄位 | 值 | 語言 |
|---|---|---|
| dc.contributor.advisor | 謝宏昀 | zh_TW |
| dc.contributor.advisor | Hung-Yun Hsieh | en |
| dc.contributor.author | 邱品萍 | zh_TW |
| dc.contributor.author | Pin-Ping Ciou | en |
| dc.date.accessioned | 2025-02-25T16:30:39Z | - |
| dc.date.available | 2026-02-11 | - |
| dc.date.copyright | 2025-02-25 | - |
| dc.date.issued | 2025 | - |
| dc.date.submitted | 2025-02-13 | - |
| dc.identifier.citation | [1] T. B. Brown, B. Mann, N. Ryder, et al., “Language models are few-shot learners,” in Proceedings of the 34th International Conference on Neural Information Processing Systems, ser. NIPS ’20, 2020.
[2] J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of deep bidirectional transformers for language understanding,” in Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Jun. 2019, pp. 4171–4186. [3] Z. Yang, Z. Dai, Y. Yang, J. Carbonell, R. Salakhutdinov, and Q. V. Le, “Xlnet: Generalized autoregressive pretraining for language understanding,” in Proceedings of the 33rd International Conference on Neural Information Processing Systems. 2019. [4] L. Huang, W. Yu, W. Ma, et al., “A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions,” ACM Trans. Inf. Syst., Nov. 2024, Just Accepted. [5] Y. Bang, S. Cahyawijaya, N. Lee, et al., “A multitask, multilingual, multimodal evaluation of ChatGPT on reasoning, hallucination, and interactivity,” in Proceedings of the 13th International Joint Conference on Natural Language Processing and the 3rd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics (Volume 1: Long Papers), Nov. 2023. [6] J. Li, X. Cheng, X. Zhao, J.-Y. Nie, and J.-R. Wen, “HaluEval: A large-scale hallucination evaluation benchmark for large language models,” in Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Dec. 2023, pp. 6449–6464. [7] S. Min, K. Krishna, X. Lyu, et al., “FActScore: Fine-grained Atomic Evaluation of Factual Precision in Long Form Text Generation,” in Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Dec. 2023, pp. 12 076–12 100. [8] S. Dhuliawala, M. Komeili, J. Xu, et al., “Chain-of-verification reduces hallucination in large language models,” in Findings of the Association for Computational Linguistics: ACL 2024, Aug. 2024, pp. 3563–3578. [9] N. Varshney, W. Yao, H. Zhang, J. Chen, and D. Yu, A stitch in time saves nine: Detecting and mitigating hallucinations of llms by validating low-confidence generation, 2023. arXiv: 2307.03987 [cs.CL]. [10] J. Luo, C. Xiao, and F. Ma, “Zero-resource hallucination prevention for large language models,” in Findings of the Association for Computational Linguistics: EMNLP 2024, Nov. 2024, pp. 3586–3602. [11] P. Manakul, A. Liusie, and M. Gales, “SelfCheckGPT: Zero-resource blackbox hallucination detection for generative large language models,” in Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Dec. 2023, pp. 9004–9017. [12] J. Zhang, Z. Li, K. Das, B. Malin, and S. Kumar, “SAC3: Reliable hallucination detection in black-box language models via semantic-aware cross-check consistency,” in Findings of the Association for Computational Linguistics: EMNLP 2023, Dec. 2023, pp. 15 445–15 458. [13] Y. Zhao, L. Yan, W. Sun, et al., “Knowing what LLMs DO NOT know: A simple yet effective self-detection method,” in Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), Jun. 2024, pp. 7051–7063. [14] M. Elaraby, M. Lu, J. Dunn, et al., Halo: Estimation and reduction of hallucinations in open-source weak large language models, 2023. arXiv: 2308. 11764 [cs.CL]. [15] T. Bever, “The cognitive basis for linguistic structures,” in Jan. 1970, pp. 279–352. [16] F. Ferreira, “The misinterpretation of noncanonical sentences,” Cognitive Psychology, vol. 47, no. 2, pp. 164–203, 2003. [17] D. N. Gunraj, A. M. Drumm-Hewitt, E. M. Dashow, S. S. N. Upadhyay, and C. M. Klin, “Texting insincerely: The role of the period in text messaging,” Computers in Human Behavior, vol. 55, pp. 1067–1075, 2016. [18] K. J. Houghton, S. S. N. Upadhyay, and C. M. Klin, “Punctuation in text messages may convey abruptness. period,” Comput. Hum. Behav., vol. 80, no. C, pp. 112–121, Mar. 2018. [19] C. Perfetti, “Reading ability: Lexical quality to comprehension,” Scientific Studies of Reading - SCI STUD READ, vol. 11, pp. 357–383, Sep. 2007. [20] J. Park, S. Min, J. Kang, L. Zettlemoyer, and H. Hajishirzi, “FaVIQ: Fact verification from information seeking questions,” in ACL, 2022. [21] A. Patel, S. Bhattamishra, and N. Goyal, “Are NLP models really able to solve simple math word problems?” In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Jun. 2021, pp. 2080–2094. [22] Z. Ji, N. Lee, R. Frieske, et al., “Survey of hallucination in natural language generation,” ACM Comput. Surv., vol. 55, no. 12, Mar. 2023. [23] T. Goyal and G. Durrett, “Annotating and Modeling Fine-grained Factuality in Summarization,” in Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Jun. 2021, pp. 1449–1462. [24] X. Lu, S. Welleck, L. Jiang, et al., “Quark: Controllable Text Generation with Reinforced Unlearning,” ArXiv, vol. abs/2205.13636, 2022. [25] S. Lin, J. Hilton, and O. Evans, “Teaching Models to Express Their Uncertainty in Words,” Transactions on Machine Learning Research, 2022. [26] I.-C. Chern, S. Chern, S. Chen, et al., FacTool: Factuality Detection in Generative AI – A Tool Augmented Framework for Multi-Task and MultiDomain Scenarios, 2023. arXiv: 2307.13528 [cs.CL]. [27] S. Huo, N. Arabzadeh, and C. L. A. Clarke, Retrieving supporting evidence for llms generated answers, 2023. arXiv: 2306.13781 [cs.IR]. [28] S. Kadavath, T. Conerly, A. Askell, et al., Language Models (Mostly) Know What They Know, 2022. arXiv: 2207.05221 [cs.CL]. [29] X. Zhang, B. Peng, Y. Tian, et al., Self-alignment for factuality: Mitigating hallucinations in llms via self-evaluation, 2024. arXiv: 2402.09267 [cs.CL]. [30] J. Li, J. Chen, R. Ren, et al., “The Dawn After the Dark: An Empirical Study on Factuality Hallucination in Large Language Models,” in Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Aug. 2024, pp. 10 879–10 899. [31] D. Zheng, M. Lapata, and J. Z. Pan, How Reliable are LLMs as Knowledge Bases? Re-thinking Facutality and Consistency, 2024. arXiv: 2407.13578 [cs.CL]. [32] J.-Y. Yao, K.-P. Ning, Z.-H. Liu, M.-N. Ning, Y.-Y. Liu, and L. Yuan, LLM Lies: Hallucinations are not Bugs, but Features as Adversarial Examples, 2024. arXiv: 2310.01469 [cs.CL]. [33] M. Xiong, Z. Hu, X. Lu, et al., Can llms express their uncertainty? an empirical evaluation of confidence elicitation in llms, 2024. arXiv: 2306.13063 [cs.CL]. [34] A. Agrawal, M. Suzgun, L. Mackey, and A. Kalai, “Do Language Models Know When They’re Hallucinating References?” In Findings of the Association for Computational Linguistics: EACL 2024, Mar. 2024, pp. 912–928. [35] R. Cohen, M. Hamri, M. Geva, and A. Globerson, “LM vs LM: Detecting Factual Errors via Cross Examination,” in Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Dec. 2023, pp. 12 621–12 640. [36] Q. Cheng, T. Sun, W. Zhang, et al., Evaluating hallucinations in chinese large language models, 2023. [37] C. Dou, Y. Zhang, Y. Chen, et al., “Detection, diagnosis, and explanation: A benchmark for Chinese medial hallucination evaluation,” in Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), May 2024. [38] X. Liang, S. Song, S. Niu, et al., “UHGEval: Benchmarking the hallucination of Chinese large language models via unconstrained generation,” in Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Aug. 2024. [39] S. Witteveen and M. Andrews, “Paraphrasing with large language models,” in Proceedings of the 3rd Workshop on Neural Generation and Translation, Nov. 2019, pp. 215–220. [40] K. R. McKeown, “Paraphrasing questions using given and new information,” American Journal of Computational Linguistics, vol. 9, no. 1, pp. 1–10, 1983. [41] D. LIN and P. PANTEL, “Discovery of inference rules for question-answering,” Natural Language Engineering, vol. 7, no. 4, pp. 343–360, 2001. [42] R. Barzilay and L. Lee, “Learning to paraphrase: An unsupervised approach using multiple-sequence alignment,” in Proceedings of the 2003 Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics, 2003, pp. 16–23. [43] I. A. Bolshakov and A. Gelbukh, “Synonymous paraphrasing using wordnet and internet,” in Lecture Notes in Computer Science. Springer Berlin Heidelberg, 2004, pp. 312–323. [44] D. Kauchak and R. Barzilay, “Paraphrasing for automatic evaluation,” in Proceedings of the Human Language Technology Conference of the NAACL, Main Conference, R. C. Moore, J. Bilmes, J. Chu-Carroll, and M. Sanderson, Eds., New York City, USA: Association for Computational Linguistics, Jun. 2006, pp. 455–462. [45] S. Wubben, A. van den Bosch, and E. Krahmer, “Paraphrase generation as monolingual translation: Data and evaluation,” in Proceedings of the 6th International Natural Language Generation Conference, Jul. 2010. [46] P. Koehn, H. Hoang, A. Birch, et al., “Moses: Open source toolkit for statistical machine translation,” in Proceedings of the 45th Annual Meeting of the Association for Computational Linguistics Companion Volume Proceedings of the Demo and Poster Sessions, Jun. 2007, pp. 177–180. [47] F. Fitrawati and D. Safitri, “Students’ grammatical errors in essay writing: A pedagogical grammar reflection,” International Journal of Language Education, vol. 5, p. 74, Jun. 2021. [48] D. P. Gadgile, “Common grammatical errors in English language committed by senior college level students,” International Journal of Science and Research (IJSR), vol. 11, no. 3, pp. 1413–1414, Mar. 2022. [49] K. M. Alghazo and M. K. Alshraideh, “Grammatical Errors Found in English Writing: A Study from Al-Hussein Bin Talal University,” International Education Studies, vol. 13, no. 9, p. 1, Aug. 2020. [50] N.-R. Han, M. Chodorow, and C. Leacock, “Detecting errors in english article usage by non-native speakers,” Natural Language Engineering, vol. 12, pp. 115–129, Jun. 2006. [51] N. Anh, N. Yen, N. Tho, and L. Nhut, “Grammatical errors in academic writing of english second-year students,” European Journal of English Language Teaching, vol. 7, Nov. 2022. [52] L. Osterhout and P. J. Holcomb, “Event-related brain potentials elicited by syntactic anomaly,” Journal of Memory and Language, vol. 31, no. 6, pp. 785–806, 1992. [53] A. Friederici, “The brain basis of language processing: From structure to function,” Physiological reviews, vol. 91, pp. 1357–92, Oct. 2011. [54] E. Nicoladis, A. Duggal, and A. Besoi, “Texting!!!: Attributions of gender and friendliness to texters who use exclamation marks,” Interaction Studies. Social Behaviour and Communication in Biological and Artificial Systems, vol. 24, pp. 422–436, Feb. 2024. [55] C. A. Perfetti, L. Hart, et al., “The lexical quality hypothesis,” Precursors of functional literacy, vol. 11, pp. 67–86, 2002. [56] C. Clifton, F. Ferreira, J. M. Henderson, et al., “Eye movements in reading and information processing: Keith Rayner’s 40year legacy,” Journal of Memory and Language, vol. 86, pp. 1–19, 2016. [57] M. A. Just and P. A. Carpenter, “A theory of reading: From eye fixations to comprehension.,” Psychological Review, vol. 87, pp. 329–354, 4 1980. [58] S. Bird, E. Klein, and E. Loper, Natural language processing with Python: analyzing text with the natural language toolkit. ” O’Reilly Media, Inc.”, 2009. [59] B. Merialdo, “Tagging English text with a probabilistic model,” Computational Linguistics, vol. 20, no. 2, J. Hirschberg, Ed., pp. 155–171, 1994. [60] A. Radford, J. Wu, R. Child, D. Luan, D. Amodei, I. Sutskever, et al., “Language models are unsupervised multitask learners,” OpenAI blog, 2019. [61] M. Honnibal and I. Montani, “spaCy 2: Natural language understanding with Bloom embeddings, convolutional neural networks and incremental parsing,” To appear, 2018. [62] J. Nivre, J. Hall, J. Nilsson, et al., “Maltparser: A language-independent system for data-driven dependency parsing,” Natural Language Engineering, vol. 13, pp. 95–135, Jun. 2007. [63] T. Kwiatkowski, J. Palomaki, O. Redfield, et al., “Natural Questions: A Benchmark for Question Answering Research,” Transactions of the Association for Computational Linguistics, vol. 7, pp. 452–466, 2019. [64] S. Min, J. Michael, H. Hajishirzi, and L. Zettlemoyer, “AmbigQA: Answering Ambiguous Open-domain Questions,” in Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Nov. 2020, pp. 5783–5797. [65] A. Abujabal, R. Saha Roy, M. Yahya, and G. Weikum, “ComQA: A Communitysourced Dataset for Complex Factoid Question Answering with Paraphrase Clusters,” in Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Jun. 2019, pp. 307–317. [66] K. Cobbe, V. Kosaraju, M. Bavarian, et al., “Training verifiers to solve math word problems,” arXiv preprint arXiv:2110.14168, 2021. | - |
| dc.identifier.uri | http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/97019 | - |
| dc.description.abstract | 大型語言模型在自然語言處理領域展現卓越性能,但它們經常產生錯誤信息,這種現象被稱為「幻覺」。現有的幻覺偵測方法面臨雙重侷限:過度仰賴外部知識庫,以及需深入模型內部結構進行分析,這對於基於API的模型來說是不切實際的。雖然近期基於一致性的方法通過問題改寫提供了一個有前景的方向,但由於它們依賴語言模型,可能會引入額外的不穩定性和潛在的幻覺。為了解決這些問題,我們提出了一種基於認知語言學的新穎問題改寫方法,消除了對語言模型的依賴。受到認知語言學研究的啟發,研究表明人類在面對認知挑戰時仍能保持語義理解,因此我們提出了假設:健全的語言模型在面對這些挑戰時也應該同樣保持語義,其中回覆的不一致性將表明潛在的幻覺。我們的方法引入了四種轉換方法:語法錯誤變化、句法結構轉換、標點符號修改及罕用詞彙替換,每種方法針對不同的認知挑戰,同時保持語義。我們在事實性問答和邏輯推理任務的數據集上評估了我們的框架。實驗結果顯示顯著改進:在事實性問答中,我們三種方法的最佳組合實現了0.755和0.720的F1分數,超過了基準的0.600和0.603分數。對於邏輯推理任務,我們的標點符號修改方法單獨就實現了0.850和0.955的F1分數,顯著優於基準的0.649和0.762分數。這些結果表明,在不需要使用額外語言模型的情況下,我們的框架提供了更可靠的幻覺檢測。 | zh_TW |
| dc.description.abstract | Large Language Models (LLMs) have demonstrated remarkable capabilities in natural language processing tasks, but they frequently generate incorrect information, a phenomenon known as hallucination. Current approaches to hallucination detection either rely heavily on external knowledge bases or require access to model internals, which is impractical for API-based models. While recent consistency-based methods offer a promising direction through question paraphrasing, they may introduce additional instability and potential hallucinations due to their dependence on language models. To address these problems, we propose a novel cognitive-linguistics-based approach for paraphrasing questions that eliminates dependence on language models.
Inspired by cognitive linguistics research showing that humans maintain semantic understanding despite cognitive challenges, we hypothesize that robust language models should similarly preserve meaning when facing these challenges, where inconsistencies in responses would indicate potential hallucinations. Our approach introduces four transformation methods: Grammatical Errors Variation, Sentence Structure Conversion, Punctuation Modification, and Lexical Substitution with Rare Words, each targeting distinct cognitive challenges while preserving meaning. We evaluate our framework on datasets covering both factoid question answering and logical reasoning tasks. Our experimental results demonstrate significant improvements: in factoid question answering, our optimal combination of three methods achieves F1 scores of 0.755 and 0.720, surpassing the baseline's scores of 0.600 and 0.603. For logical reasoning tasks, our Punctuation Modification method alone achieves F1 scores of 0.850 and 0.955, significantly outperforming the baseline's scores of 0.649 and 0.762. These results demonstrate that our framework provides more reliable and efficient hallucination detection by eliminating multiple language model calls. | en |
| dc.description.provenance | Submitted by admin ntu (admin@lib.ntu.edu.tw) on 2025-02-25T16:30:39Z No. of bitstreams: 0 | en |
| dc.description.provenance | Made available in DSpace on 2025-02-25T16:30:39Z (GMT). No. of bitstreams: 0 | en |
| dc.description.tableofcontents | ABSTRACT. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ii
LIST OF TABLES. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vi LIST OF FIGURES. . . . . . . . . . . . . . . . . . . . . . . . . . . . . viii CHAPTER 1 INTRODUCTION. . . . . . . . . . . . . . . . . . . . 1 CHAPTER 2 BACKGROUND AND RELATED WORK. . . . . 4 2.1 Large Language Models (LLMs) . . . . . . . . . . . . . . . . . . . 4 2.2 Hallucination in Large Language Models . . . . . . . . . . . . . . 4 2.3 Hallucination Detection for Large Language Models . . . . . . . . 5 2.3.1 Fact-checking Approaches . . . . . . . . . . . . . . . . . . . 6 2.3.2 Uncertainty Estimation Methods . . . . . . . . . . . . . . . 6 2.4 Hallucination Detection in Chinese Language Models . . . . . . . . 7 2.5 Paraphrasing in Questions Generation . . . . . . . . . . . . . . . . 8 2.5.1 Language Model Dependency in Current Methods . . . . . 8 2.5.2 Traditional Paraphrasing Approaches . . . . . . . . . . . . 9 2.5.3 Limitations of Traditional Paraphrasing Approaches . . . . 10 2.6 Cognitive Linguistic Foundations of Processing Challenges . . . . . 10 2.6.1 Cognitive Linguistics Foundation . . . . . . . . . . . . . . . 10 2.6.2 Cognitive Challenges in Grammatical Errors . . . . . . . . 12 2.6.3 Cognitive Challenges in Sentence Structure Processing . . . 13 2.6.4 Cognitive Challenges in Punctuation . . . . . . . . . . . . . 14 2.6.5 Cognitive Challenges in Low-frequency Words . . . . . . . 14 CHAPTER 3 SYSTEM MODEL. . . . . . . . . . . . . . . . . . . . 16 3.1 Motivation and Overview . . . . . . . . . . . . . . . . . . . . . . . 16 3.1.1 Limitations of Existing Prompt-Based Approaches . . . . . 16 3.1.2 Cognitive-Linguistics-Based Approach . . . . . . . . . . . . 16 3.2 System Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 3.3 Question Generation . . . . . . . . . . . . . . . . . . . . . . . . . . 17 3.4 Large Language Models (LLMs) Generation . . . . . . . . . . . . . 18 3.5 Consistency Detection . . . . . . . . . . . . . . . . . . . . . . . . . 18 CHAPTER 4 COGNITIVE-LINGUISTICS-BASED QUESTION PARAPHRASING MODULE. . . . . . . . . . . . . . . . . . . . . 21 4.1 Module Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . 21 4.2 Grammatical Errors Variation . . . . . . . . . . . . . . . . . . . . 22 4.2.1 Motivation and Theoretical Foundation . . . . . . . . . . . 22 4.2.2 Method Design . . . . . . . . . . . . . . . . . . . . . . . . . 22 4.3 Sentence Structure Conversion . . . . . . . . . . . . . . . . . . . . 27 4.3.1 Motivation and Theoretical Foundation . . . . . . . . . . . 27 4.3.2 Method Design . . . . . . . . . . . . . . . . . . . . . . . . . 27 4.4 Punctuation Modification . . . . . . . . . . . . . . . . . . . . . . . 29 4.4.1 Motivation and Theoretical Foundation . . . . . . . . . . . 29 4.4.2 Method Design . . . . . . . . . . . . . . . . . . . . . . . . . 31 4.5 Lexical Substitution with Rare Words . . . . . . . . . . . . . . . . 32 4.5.1 Motivation and Theoretical Foundation . . . . . . . . . . . 32 4.5.2 Method Design . . . . . . . . . . . . . . . . . . . . . . . . . 33 4.6 Experimental Design and Evaluation Strategy . . . . . . . . . . . 36 4.6.1 Method Notation . . . . . . . . . . . . . . . . . . . . . . . 36 4.6.2 Experimental Flow and Evaluation Process . . . . . . . . . 36 CHAPTER 5 EVALUATION AND ANALYSIS. . . . . . . . . . . 38 5.1 Dataset . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38 5.1.1 Factoid Question Answering . . . . . . . . . . . . . . . . . 38 5.1.2 Logical Reasoning . . . . . . . . . . . . . . . . . . . . . . . 39 5.2 Experiment Setup . . . . . . . . . . . . . . . . . . . . . . . . . . . 41 5.2.1 Ground Truth Labeling Process and Results . . . . . . . . 41 5.2.2 Baseline Method . . . . . . . . . . . . . . . . . . . . . . . . 43 5.2.3 Evaluation Metrics . . . . . . . . . . . . . . . . . . . . . . 44 5.3 Optimal Combination Analysis on Factoid Question Answering . . 46 5.3.1 Single Variant Method Performance . . . . . . . . . . . . . 46 5.3.2 Method Combination Analysis . . . . . . . . . . . . . . . . 49 5.3.3 Optimal Strategy for Factoid Question Answering . . . . . 55 5.4 Optimal Combination Analysis on Logical Reasoning . . . . . . . . 57 5.4.1 Single Variant Method Performance . . . . . . . . . . . . . 57 5.4.2 Method Combination Analysis . . . . . . . . . . . . . . . . 60 5.4.3 Optimal Strategy for Logical Reasoning . . . . . . . . . . . 61 5.5 Baseline Comparison . . . . . . . . . . . . . . . . . . . . . . . . . 63 5.5.1 Performance Improvements on Factoid Question Answering 63 5.5.2 Performance Improvements on Logical Reasoning . . . . . . 66 5.5.3 Summary of Improvements . . . . . . . . . . . . . . . . . . 68 CHAPTER 6 CONCLUSION AND FUTURE WORK. . . . . . 69 6.1 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69 6.2 Future Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69 REFERENCES. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71 | - |
| dc.language.iso | en | - |
| dc.subject | 認知挑戰 | zh_TW |
| dc.subject | 問題改寫 | zh_TW |
| dc.subject | 認知語言學 | zh_TW |
| dc.subject | 大型語言模型 | zh_TW |
| dc.subject | 幻覺檢測 | zh_TW |
| dc.subject | Cognitive Challenges | en |
| dc.subject | Large Language Models (LLMs) | en |
| dc.subject | Hallucination Detection | en |
| dc.subject | Cognitive Linguistics | en |
| dc.subject | Question Paraphrasing | en |
| dc.title | 運用認知語言學增強大型語言模型之幻覺偵測 | zh_TW |
| dc.title | Enhancing Hallucination Detection in Large Language Models through Question Paraphrasing Strategies Inspired by Cognitive Linguistics | en |
| dc.type | Thesis | - |
| dc.date.schoolyear | 113-1 | - |
| dc.description.degree | 碩士 | - |
| dc.contributor.oralexamcommittee | 高榮鴻;沈上翔;葉佳宜 | zh_TW |
| dc.contributor.oralexamcommittee | Rung-Hung Gau;Shan-Hsiang Shen;Chia-Yi Yeh | en |
| dc.subject.keyword | 大型語言模型,幻覺檢測,認知語言學,問題改寫,認知挑戰, | zh_TW |
| dc.subject.keyword | Large Language Models (LLMs),Hallucination Detection,Cognitive Linguistics,Question Paraphrasing,Cognitive Challenges, | en |
| dc.relation.page | 76 | - |
| dc.identifier.doi | 10.6342/NTU202500638 | - |
| dc.rights.note | 未授權 | - |
| dc.date.accepted | 2025-02-13 | - |
| dc.contributor.author-college | 電機資訊學院 | - |
| dc.contributor.author-dept | 電機工程學系 | - |
| dc.date.embargo-lift | N/A | - |
| 顯示於系所單位: | 電機工程學系 | |
文件中的檔案:
| 檔案 | 大小 | 格式 | |
|---|---|---|---|
| ntu-113-1.pdf 未授權公開取用 | 3.23 MB | Adobe PDF |
系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。
