Skip navigation

DSpace JSPUI

DSpace preserves and enables easy and open access to all types of digital content including text, images, moving images, mpegs and data sets

Learn More
DSpace logo
English
中文
  • Browse
    • Communities
      & Collections
    • Publication Year
    • Author
    • Title
    • Subject
    • Advisor
  • Search TDR
  • Rights Q&A
    • My Page
    • Receive email
      updates
    • Edit Profile
  1. NTU Theses and Dissertations Repository
  2. 電機資訊學院
  3. 資訊工程學系
Please use this identifier to cite or link to this item: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/96041
Title: 基於可學習編碼簿之大型語言模型提示詞微調
Prompt Tuning of Large Language Models Based on Learnable Codebooks
Authors: 林昱辰
Yu-Chen Lin
Advisor: 陳祝嵩
Chu-Song Chen
Keyword: 輕量化微調,提示詞微調,乘積量化,深度學習,遷移式學習,
Parameter Efficient Fine-tuning,Prompt Tuning,Product Quantization,Deep Learning,Transfer Learning,
Publication Year : 2024
Degree: 碩士
Abstract: 提示詞微調是一種熱門的輕量化微調架構,其特點是在預訓練模型的基礎上僅更新少量參數,即可達到亮眼的表現。在過去的方法中,每個提示詞通常被視為一個整體,且各自獨立更新,導致隨著提示詞的增多,更新的參數量也線性增長。為了解決此問題,我們提出了高效率提示詞微調的可適應性編碼簿。我們利用了乘積量化的概念,使提示詞在每個分割後的子空間中共享一組可學習的編碼向量。每個提示詞透過一組自適應權重而有所變化。我們在 17 個自然語言任務中,僅更新預訓練模型 0.3% 的參數,就達到了優異的表現,包括自然語言理解及問答任務。此外,我們的方法在少樣本學習情境以及大型語言模型骨幹下也有良好表現,凸顯了其適應性及可發展性。
Prompt Tuning has emerged as a popular Parameter-Efficient Fine-Tuning method attributed to its excellent performance with few updated parameters on various large-scale Pretrained Language Models (PLMs). In previous approaches, each prompt has been considered as a whole and updated independently, causing all parameters to depend on prompt length and increase accordingly. To alleviate this problem, we introduce Adaptive Codebook for Composite and Efficient Prompt Tuning (ACCEPT). In our approach, we utilize the concept of product quantization (PQ), enabling all soft prompts to share a common set of learnable codebook vectors within each subspace. Each prompt is then distinguished by a unique set of adaptive weights. We achieve impressive performances on 17 diverse natural language tasks, including natural language understanding (NLU) and question answering (QA), by training only 0.3% of parameters of the PLMs. Additionally, our method excels in fewshot and large model scenarios, highlighting its significant adaptability and potential.
URI: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/96041
DOI: 10.6342/NTU202402283
Fulltext Rights: 同意授權(限校園內公開)
metadata.dc.date.embargo-lift: 2025-09-01
Appears in Collections:資訊工程學系

Files in This Item:
File SizeFormat 
ntu-112-2.pdf
Access limited in NTU ip range
604.8 kBAdobe PDF
Show full item record


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

社群連結
聯絡資訊
10617臺北市大安區羅斯福路四段1號
No.1 Sec.4, Roosevelt Rd., Taipei, Taiwan, R.O.C. 106
Tel: (02)33662353
Email: ntuetds@ntu.edu.tw
意見箱
相關連結
館藏目錄
國內圖書館整合查詢 MetaCat
臺大學術典藏 NTU Scholars
臺大圖書館數位典藏館
本站聲明
© NTU Library All Rights Reserved