請用此 Handle URI 來引用此文件:
http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/96077
標題: | 使用PEFT在維持高效率的訓練下對使用者評論進行基於面向的情感分析 ABSA : Opinion Tree Parsing with PEFT for Aspect-based Semtiment Analysis |
作者: | 王敬順 Ching-Shun Wang |
指導教授: | 廖世偉 Shih-Wei Liao |
關鍵字: | 主題提取,主題意見提取,情感分析,自然語言處理,社群媒體使用者留言, Aspect Extract,Opinion Extract,Aspect-based Sentiment Analysis,Natural Language Processing,Comments from social media users, |
出版年 : | 2024 |
學位: | 碩士 |
摘要: | 分析社交媒體用戶的評論呈現出重大挑戰,因為要辨識留言主體與留言情感之間的關係在複雜性上有所困難,特別是當使用者的評論在長度上有很大的變化時。本文介紹了一種新穎的意見樹解析模型,該模型能夠處理評論中不同方面之間錯綜複雜的互動,在訓練模型時,加入連接詞和語義修飾詞來提高解析的準確度。且在模型複雜化之下,為了提高訓練過程的效率並管理計算需求,我們在模型上實作了可以使參數量減少卻能達到差不多效能的方法(PEFT)。
我們在 ACOS 數據集上評估了我們提出的模型,鑑於描述用戶對特定方面情感的數據集有限,以及由於其資源密集性對大型預訓練語言模型(LLMs)進行下游調整的挑戰,我們的方法提出了一種改變計算方式的 OTP 模型。這種方法改變了模型的Loss function,專注於戰略性放置的模塊訓練,且在加入Adpater的情況下,顯著減少了 GPU 記憶體占用,並減輕了記憶體不足(OOM)問題,而不損害預訓練模型的整體完整性。這種方法不僅提高了訓練效率,而且還維持了與原始 LLM 配置接近的性能水平。 Analyzing social media user comments presents significant challenges due to the complexity of discerning relationships between opinions and aspects, particularly when comments vary greatly in length. This paper introduces a novel Opinion Tree Parser Model that navigates the intricate interplay between different aspects within comments, utilizing conjunctions and semantic modifiers to enhance the parsing accuracy. To improve the efficiency of the training process and manage the computational demands, we have implemented Position-Encoded Fine-Tuning (PEFT) methods on the decoder side. We evaluated our proposed model on ACOS datasets, given the limited availability of datasets that describe user sentiments towards specific aspects and the challenges of fine-tuning large pre-trained language models (LLMs) due to their resource intensity, our approach proposes an advanced context-free opinion grammar. This method integrates an adapter to focus training on strategically placed modules, significantly reducing the GPU memory footprint and mitigating out-of-memory (OOM) issues without compromising the overall integrity of the pre-trained model. This approach not only enhances training efficiency but also maintains performance levels close to those of the original LLM configurations. |
URI: | http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/96077 |
DOI: | 10.6342/NTU202404405 |
全文授權: | 未授權 |
顯示於系所單位: | 資訊工程學系 |
文件中的檔案:
檔案 | 大小 | 格式 | |
---|---|---|---|
ntu-113-1.pdf 目前未授權公開取用 | 712.84 kB | Adobe PDF |
系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。