Skip navigation

DSpace JSPUI

DSpace preserves and enables easy and open access to all types of digital content including text, images, moving images, mpegs and data sets

Learn More
DSpace logo
English
中文
  • Browse
    • Communities
      & Collections
    • Publication Year
    • Author
    • Title
    • Subject
    • Advisor
  • Search TDR
  • Rights Q&A
    • My Page
    • Receive email
      updates
    • Edit Profile
  1. NTU Theses and Dissertations Repository
  2. 電機資訊學院
  3. 資訊工程學系
Please use this identifier to cite or link to this item: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/101178
Full metadata record
???org.dspace.app.webui.jsptag.ItemTag.dcfield???ValueLanguage
dc.contributor.advisor吳家麟zh_TW
dc.contributor.advisorJa-Ling Wuen
dc.contributor.author劉品均zh_TW
dc.contributor.authorPin-Jiun Liuen
dc.date.accessioned2025-12-31T16:13:23Z-
dc.date.available2026-01-01-
dc.date.copyright2025-12-31-
dc.date.issued2025-
dc.date.submitted2025-12-08-
dc.identifier.citation[1] Gregory K. Wallace. The JPEG still picture compression standard. Communications of the ACM, 34(4):30–44, 1991.
[2] David S. Taubman and Michael W. Marcellin. JPEG2000: Image Compression Fundamentals, Standards and Practice. Springer Science & Business Media, 2002.
[3] Fabrice Bellard. BPG image format. http://bellard.org/bpg/. Accessed: 2025-07-01.
[4] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems, 25:1097–1105, 2012.
[5] Johannes Ballé, Valero Laparra, and Eero P. Simoncelli. End-to-end optimized image compression. In International Conference on Learning Representations (ICLR), 2017. arXiv:1611.01704.
[6] Johannes Ballé, David Minnen, Saurabh Singh, Sung Jin Hwang Johnston, and George Toderici. Variational image compression with a scale hyperprior. In ICLR, 2018.
[7] Zhengxue Cheng, Heming Sun, Masaru Takeuchi, and Jiro Katto. Learned image compression with discretized Gaussian mixture likelihoods and attention modules. In CVPR, pages 7939–7948, 2020. doi:10.6342/NTU202501437
[8] Ashish Vaswani et al. Attention is all you need. Advances in Neural Information Processing Systems, 30, 2017.
[9] Alexey Dosovitskiy et al. An image is worth 16×16 words: Transformers for image recognition at scale. ICLR, 2021.
[10] Ze Liu et al. Swin Transformer: Hierarchical vision transformer using shifted windows. In ICCV, pages 10012–10022, 2021.
[11] Renjie Zou et al. Swin-transformer based image compression with large receptive field and high adaptability. In CVPR Workshops, 2022.
[12] Jinming Liu et al. Learned image compression with mixed transformer–CNN architectures. In CVPR, 2023.
[13] 許興宇. 透過漸進式學習實現基於 LoRA 的可變壓縮率深度影像壓縮. 臺灣大學資訊網路與多媒體研究所碩士論文, 2024. doi:10.6342/NTU202501437
[14] Edward J. Hu et al. LoRA: Low-rank adaptation of large language models. ICLR, 2022.
[15] Tim Dettmers et al. QLoRA: Efficient finetuning of quantized LLMs. arXiv:2305.14314, 2023.
[16] Yitao Zhang et al. AdaLoRA: Adaptive low-rank adaptation for efficient fine-tuning. arXiv:2303.10512, 2023.
[17] Haotian Peng et al. AsyLoRA: Asymmetric low-rank adaptation for vision-and-language pre-training. arXiv:2309.16736, 2023.
[18] Shizhe Gu, Ziwei Liu, and Dahua Lin. DoRA: Weight-decomposed low-rank adaptation. arXiv:2402.09353, 2024.
[19] David Minnen, Johannes Ballé, and George Toderici. Joint autoregressive and hierarchical priors for learned image compression. In NeurIPS, pages 10771–10780, 2018.
[20] Gisle Bjontegaard. Calculation of average PSNR differences between RD-curves. Technical Report VCEG-M33, ITU-T Video Coding Experts Group, 2001. https://www.itu.int/md/T01-SG16-030409-TD-GEN-0003/en
-
dc.identifier.urihttp://tdr.lib.ntu.edu.tw/jspui/handle/123456789/101178-
dc.description.abstract隨著深度學習技術的發展,深度影像壓縮模型已大量應用於日常生活中。然而, 傳統完整微調 (Full Fine-Tuning, FT)方式需耗費大量時間與計算資源,限制了可變率影像壓縮 (Variable-Rate Image Compression) 技術在資源受限設備上的應用。為此,本研究以系統性方式探討參數高效微調技術 LoRA (Low-Rank Adaptation) 及其變形體 (AsyLoRA 與 DoRA) 在可變率影像壓縮中的應用效能。 實驗結果顯示,LoRA 及其變形體相較於傳統完整微調方式,能將需更新參數量降至原先需求約 1% 以下,並將記憶體儲存需求從原先的 877 MB 降至僅約 4 MB,同時還能維持良好的壓縮功效。特別是 DoRA,在僅略增儲存開銷前提下,提供比 LoRA 更佳且更穩定的效能; 而 AsyLoRA 雖然相較於DoRA效能稍低,但所需記憶體量最低,特別適合應用於極低資源的環境。 總結而言,這些低秩調適方法在可變率影像壓縮中展現出高度的應用潛力,本研究為未來在資源受限情境下可變率影像壓縮的應用奠定了初步的基礎zh_TW
dc.description.abstractWith the development of deep learning technologies, deep image compression models have been widely applied in daily life. However, traditional full fine-tuning (FT) requires significant time and computational resources, limiting the practical deployment of variable-rate image compression techniques on resource-constrained devices. To address this, this study systematically investigates the effectiveness of parameter-efficient fine-tuning techniques, namely LoRA (Low-Rank Adaptation) and its variants (AsyLoRA and DoRA), in variable-rate image compression.
Experimental results show that LoRA and its variants can reduce the number of trainable parameters to about 1% of that required by traditional FT and lower storage requirements from 877 MB to approximately 4 MB, while still maintaining good compression performance. In particular, DoRA offers better and more stable performance than LoRA with only a slight increase in storage overhead. Although AsyLoRA shows slightly lower performance compared to DoRA, it achieves the lowest memory consumption, making it especially suitable for extremely resource-limited scenarios.
In summary, these low-rank adaptation methods exhibit great potential for variable-rate image compression and lay a solid foundation for future applications on resource-constrained devices.
en
dc.description.provenanceSubmitted by admin ntu (admin@lib.ntu.edu.tw) on 2025-12-31T16:13:23Z
No. of bitstreams: 0
en
dc.description.provenanceMade available in DSpace on 2025-12-31T16:13:23Z (GMT). No. of bitstreams: 0en
dc.description.tableofcontents摘要 i
Abstract iii
Contents v
List of Figures vii
List of Tables ix
Chapter 1 Introduction 1
1.1 Image Compression 1
1.2 Variable-Rate Image Compression 2
1.3 Motivation 5
Chapter 2 Related Work 9
2.1 LoRA 9
2.2 Asymmetry LoRA 10
2.3 DoRA 11
2.4 Learned Image Compression with Mixed Transformer-CNN Architectures 14
Chapter 3 Proposed Method 19
3.1 Experiment Setup 19
3.2 Datasets 20
3.3 Performance Evaluation Indices 21
Chapter 4 Experiment Ⅰ 25
4.1 Anchor Point Selection 25
4.2 Layer Selection 28
4.3 Different Rank r for LoRA 29
Chapter 5 Experiment Ⅱ 31
5.1 Asymmetry LoRA's Results and Discussion 31
5.2 DoRA's Results and Discussion 34
Chapter 6 LoRA and Its Variants on Deep Image Compression Discussion 37
6.1 Computation-Efficiency Comparison 37
6.2 Memory-Efficient Comparison 39
Chapter 7 Conclusions 41
References 43
-
dc.language.isoen-
dc.subject深度影像壓縮-
dc.subject可變率影像壓縮-
dc.subject低秩適應-
dc.subject非對稱低秩適應-
dc.subject權重分解式低秩適應-
dc.subjectDeep image compression-
dc.subjectVariable rate image compression-
dc.subjectLoRA-
dc.subjectAsyLoRA-
dc.subjectDoRA-
dc.titleLoRA 與其變體應用於深度影像壓縮模型之效益探討zh_TW
dc.titleEffectiveness of LoRA and Its Variants on Deep Image Compression Modelsen
dc.typeThesis-
dc.date.schoolyear114-1-
dc.description.degree碩士-
dc.contributor.oralexamcommittee陳駿丞;李明穗zh_TW
dc.contributor.oralexamcommitteeJun-Cheng Chen;Ming-Sui Leeen
dc.subject.keyword深度影像壓縮,可變率影像壓縮低秩適應非對稱低秩適應權重分解式低秩適應zh_TW
dc.subject.keywordDeep image compression,Variable rate image compressionLoRAAsyLoRADoRAen
dc.relation.page45-
dc.identifier.doi10.6342/NTU202501437-
dc.rights.note未授權-
dc.date.accepted2025-12-08-
dc.contributor.author-college電機資訊學院-
dc.contributor.author-dept資訊工程學系-
dc.date.embargo-liftN/A-
Appears in Collections:資訊工程學系

Files in This Item:
File SizeFormat 
ntu-114-1.pdf
  Restricted Access
3.59 MBAdobe PDF
Show simple item record


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

社群連結
聯絡資訊
10617臺北市大安區羅斯福路四段1號
No.1 Sec.4, Roosevelt Rd., Taipei, Taiwan, R.O.C. 106
Tel: (02)33662353
Email: ntuetds@ntu.edu.tw
意見箱
相關連結
館藏目錄
國內圖書館整合查詢 MetaCat
臺大學術典藏 NTU Scholars
臺大圖書館數位典藏館
本站聲明
© NTU Library All Rights Reserved