Skip navigation

DSpace JSPUI

DSpace preserves and enables easy and open access to all types of digital content including text, images, moving images, mpegs and data sets

Learn More
DSpace logo
English
中文
  • Browse
    • Communities
      & Collections
    • Publication Year
    • Author
    • Title
    • Subject
    • Advisor
  • Search TDR
  • Rights Q&A
    • My Page
    • Receive email
      updates
    • Edit Profile
  1. NTU Theses and Dissertations Repository
  2. 重點科技研究學院
  3. 積體電路設計與自動化學位學程
Please use this identifier to cite or link to this item: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/99508
Title: 結合分數階高斯濾波器與剪枝之深度神經網路壓縮方法
FGFP: A Fractional Gaussian Filter and Pruning for Deep Neural Networks Compression
Authors: 杜冠廷
Kuan-Ting Tu
Advisor: 簡韶逸
Shao-Yi Chien
Keyword: 網路壓縮,分數階導數,Grünwald-Letnikov 分數階導數,高斯函數,自適應非結構化剪枝,
Networks Compression,Fractional-Order Derivative,Grünwald-Letnikov fractional derivatives,Gaussian Function,Adaptive Unstructured Pruning,
Publication Year : 2025
Degree: 碩士
Abstract: 隨著深度神經網路在實際應用中的廣泛使用,神經網路壓縮技術的重要性日益提升,特別是在資源受限的邊緣設備上。因為在現實世界的應用中,深度神經網路(DNN)往往對其造成沉重的硬體負擔。

儘管已有諸多方法致力於壓縮模型參數,部署這些模型在邊緣設備上仍面臨諸多挑戰。為了應對這個問題,我們提出了結合分數階高斯濾波器和剪枝的壓縮架構FGFP(Fractional Gaussian Filter and Pruning)。該框架整合了分數階微分和高斯函數以構建分數階高斯濾波器(FGFs)。為了降低分數階微分運算的計算複雜度,我們引入了Grünwald–Letnikov分數階導數來近似分數階微分方程式。每個FGF捲積核的參數量被精簡至僅剩七個。進一步地,為了優化運算效率,我們透過運算順序重排,在分數高斯濾波器架構中實現了卷積運算計算量減少64.06%。除分數高斯濾波器架構外,我們的FGFP架構還整合了自適應非結構化剪枝(AUP)以實現更高的壓縮率。

在各種架構和基準數據集上的實驗表明,我們的FGFP架構在準確性和壓縮率方面都優於最近的方法。在CIFAR-10上,ResNet-20在模型尺寸縮小85.2%的情況下,準確度僅下降了1.52%。在ILSVRC2012上,ResNet-50在模型尺寸縮小69.1%的情況下,準確度僅下降了1.63%。

為了評估該框架的實際適用性,我們在消費級筆記型電腦上部署了該模型進行推理。實驗結果顯示,該方法在端到端推理過程中實現了3.33倍的速度提升。
Neural network compression techniques are becoming increasingly important nowadays because Deep Neural Networks (DNNs) impose heavy hardware resource loads on edge devices in real-world applications.

While many methods compress neural network parameters, deploying these models on edge devices remains challenging. To address this, we propose the Fractional Gaussian Filter and Pruning (FGFP) framework, which integrates fractional-order differential calculus and Gaussian function to construct fractional Gaussian filters (FGFs). To reduce the computational complexity of fractional-order differential operations, we introduce Grünwald-Letnikov fractional derivatives to approximate the fractional-order differential equation. The number of parameters for each kernel in FGF is minimized to only seven. Furthermore, through operation reordering, we achieve a 64.06% reduction in the computational load of convolution operations within the fractional Gaussian filter architecture. Beyond the architecture of fractional Gaussian filters, our FGFP framework also incorporates Adaptive Unstructured Pruning (AUP) to achieve higher compression ratios.

Experiments on various architectures and benchmarks show that our FGFP framework outperforms recent methods in accuracy and compression. On CIFAR-10, ResNet-20 achieves only a 1.52% drop in accuracy while reducing the model size by 85.2%. On ILSVRC2012, ResNet-50 achieves only a 1.63% drop in accuracy while reducing the model size by 69.1%.

To assess the framework’s real-world applicability, we deployed the model on a consumer laptop, achieving a 3.33× end-to-end speedup.
URI: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/99508
DOI: 10.6342/NTU202501353
Fulltext Rights: 同意授權(全球公開)
metadata.dc.date.embargo-lift: 2030-06-26
Appears in Collections:積體電路設計與自動化學位學程

Files in This Item:
File SizeFormat 
ntu-113-2.pdf
  Until 2030-06-26
3.81 MBAdobe PDF
Show full item record


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

社群連結
聯絡資訊
10617臺北市大安區羅斯福路四段1號
No.1 Sec.4, Roosevelt Rd., Taipei, Taiwan, R.O.C. 106
Tel: (02)33662353
Email: ntuetds@ntu.edu.tw
意見箱
相關連結
館藏目錄
國內圖書館整合查詢 MetaCat
臺大學術典藏 NTU Scholars
臺大圖書館數位典藏館
本站聲明
© NTU Library All Rights Reserved