Skip navigation

DSpace

機構典藏 DSpace 系統致力於保存各式數位資料(如:文字、圖片、PDF)並使其易於取用。

點此認識 DSpace
DSpace logo
English
中文
  • 瀏覽論文
    • 校院系所
    • 出版年
    • 作者
    • 標題
    • 關鍵字
    • 指導教授
  • 搜尋 TDR
  • 授權 Q&A
    • 我的頁面
    • 接受 E-mail 通知
    • 編輯個人資料
  1. NTU Theses and Dissertations Repository
  2. 電機資訊學院
  3. 電機工程學系
請用此 Handle URI 來引用此文件: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/19554
完整後設資料紀錄
DC 欄位值語言
dc.contributor.advisor于天立(Tian-Li Yu)
dc.contributor.authorSong-Bo Yangen
dc.contributor.author楊淞博zh_TW
dc.date.accessioned2021-06-08T02:05:10Z-
dc.date.copyright2020-08-24
dc.date.issued2020
dc.date.submitted2020-08-18
dc.identifier.citation[1] D. Berthelot, N. Carlini, I. Goodfellow, N. Papernot, A. Oliver, and C. A. Raffel. Mixmatch: A holistic approach to semi-supervised learning. In Advances in Neural Information Processing Systems, pages 5050-5060, 2019.
[2] P. Cascante-Bonilla, F. Tan, Y. Qi, and V. Ordonez. Curriculum labeling: self-paced pseudo-labeling for semi-supervised learning. arXiv preprint arXiv:2001.06001, 2020.
[3] O. Chapelle, B. Scholkopf, and A. Zien. Semi-supervised learning (Chapelle, o. et al., eds.; 2006)[book reviews]. IEEE Transactions on Neural Networks, 20(3):542{542, 2009.
[4] E. D. Cubuk, B. Zoph, J. Shlens, and Q. V. Le. Randaugment: Practical data augmentation with no separate search. arXiv preprint arXiv:1909.13719, 2019.
[5] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pages 248-255. IEEE, 2009.
[6] T. DeVries and G. W. Taylor. Dataset augmentation in feature space. arXiv preprint arXiv:1702.05538, 2017.
[7] C. Doersch, A. Gupta, and A. A. Efros. Unsupervised visual representation learning by context prediction. In Proceedings of the IEEE International Conference on Computer Vision, pages 1422-1430, 2015.
[8] S. A. Everingham, Markand Eslami, L. Van Gool, C. K. Williams, J. Winn, and A. Zisserman. The pascal visual object classes challenge: A retrospective. International Journal of Computer Vision, 111(1):98-136, 2015.
[9] S. Gidaris, P. Singh, and N. Komodakis. Unsupervised representation learning by predicting image rotations. arXiv preprint arXiv:1803.07728, 2018.
[10] D. P. Kingma and M. Welling. Auto-encoding variational Bayes. arXiv preprint arXiv:1312.6114, 2013.
[11] A. Krogh and J. A. Hertz. A simple weight decay can improve generalization. In Advances in Neural Information Processing Systems, pages 950-957, 1992.
[12] D.-H. Lee. Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In Workshop on challenges in representation learning, ICML, volume 3, page 2, 2013.
[13] I. Loshchilov and F. Hutter. Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101, 2017.
[14] I. Loshchilov and F. Hutter. Fixing weight decay regularization in Adam. 2018.
[15] T. Miyato, S.-I. Maeda, M. Koyama, and S. Ishii. Virtual adversarial training: a regularization method for supervised and semi-supervised learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 41(8):1979-1993, 2018.
[16] K. Nakamura and B.-W. Hong. Adaptive weight decay for deep neural networks. IEEE Access, 7:118857{118865, 2019.
[17] J. Pont-Tuset, F. Perazzi, S. Caelles, P. Arbel´aez, A. Sorkine-Hornung, and L. Van Gool. The 2017 davis challenge on video object segmentation. arXiv preprint arXiv:1704.00675, 2017.
[18] F. Schroff, D. Kalenichenko, and J. Philbin. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 815–823, 2015.
[19] A. Tarvainen and H. Valpola. Mean teachers are better role models: Weightaveraged consistency targets improve semi-supervised deep learning results. In Advances in Neural Information Processing Systems, pages 1195–1204, 2017.
[20] X. Wang, D. Kihara, J. Luo, and G.-J. Qi. Enaet: Self-trained ensemble autoencoding transformations for semi-supervised learning. arXiv preprint arXiv:1911.09265, 2019.
[21] Q. Xie, Z. Dai, E. Hovy, M.-T. Luong, and Q. V. Le. Unsupervised data augmentation. arXiv preprint arXiv:1904.12848, 2019.
[22] S. Zagoruyko and N. Komodakis. Wide residual networks. arXiv preprint arXiv:1605.07146, 2016.
[23] H. Zhang, M. Cisse, Y. N. Dauphin, and D. Lopez-Paz. Mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412, 2017.
[24] L. Zhang, G.-J. Qi, L. Wang, and J. Luo. Aet vs. Aed : Unsupervised representation learning by auto-encoding transformations rather than data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2547–2555, 2019.
[25] R. Zhang, P. Isola, and A. A. Efros. Colorful image colorization. In European Conference on Computer Vision, pages 649–666. Springer, 2016.
dc.identifier.urihttp://tdr.lib.ntu.edu.tw/jspui/handle/123456789/19554-
dc.description.abstract近年來,半監督學習的相關技術大大降低了模型對大量標記數據的依賴,並且在ImageNet、CIFAR-10等著名的數據集上展現了令人驚豔的成果。然而因為資料集型態的差異性太大,造成許多著名的方法(尤其是基於數據增強技術的方法)在工業應用類型的資料集上仍然難以獲得顯著的改善。因此本論文的目標是在工業類型的影像資料上透過半監督學習框架來改進現有的固定架構模型,並且提高模型的準確度。
本論文提出了偽表示標記,一個簡單而靈活的框架,透過批次增加標記數據並結合自監督學習來試圖改進上述的問題。然而,在添加標記數據的同時也會產生一定程度的雜訊,進而影響模型的表現,因此我們提出了方法來衡量資料雜訊對整體準確度的影響。此外為了最小化產生的雜訊,我們傾向於小批次的加入標記數據,並且會對批次量的大小進行探討。相較於其他的半監督學習演算法,這個框架更為彈性,可以根據不同的資料型態來選擇自監督學習的任務,並在大量的未標籤資料上學習到整個資料集的特性。在我們的實驗中,在工業類型的分類問題(例如WM-811K晶圓圖和MIT-BIH 心律不整分類)上,此框架的表現優於當前的最新半監督學習方法。
zh_TW
dc.description.abstractIn recent years, semi-supervised learning (SSL) has shown tremendous success in leveraging unlabeled data to improve the performance of deep learning models, which significantly reduces the demand for large amounts of labeled data. Many SSL techniques have been proposed and have shown promising performance on famous datasets such as ImageNet and CIFAR-10. However, some exiting techniques (especially data augmentation based) are not suitable for industrial applications empirically. Therefore, this work proposes the pseudo-representation labeling, a simple and flexible framework that utilizes pseudo-labeling techniques to iteratively label a small amount of unlabeled data and use them as training data. In addition, our framework is integrated with self-supervised representation learning such that the classifier gains benefits from both labeled and unlabeled data. This framework can be implemented without being limited at the specific model structure, but a general technique to improve the existing model. Compared with the existing approaches, the pseudo-representation labeling is more intuitive and can effectively solve practical problems in the real world. Empirically, it outperforms the current state-of-the-art semi-supervised learning methods in industrial types of classification problems such as the WM-811K wafer map and the MIT-BIH Arrhythmia dataset.en
dc.description.provenanceMade available in DSpace on 2021-06-08T02:05:10Z (GMT). No. of bitstreams: 1
U0001-1608202013510800.pdf: 14609250 bytes, checksum: fbe999fb24da1decff47eecddc322a70 (MD5)
Previous issue date: 2020
en
dc.description.tableofcontents口試委員會審定書 . . . . . . . . . . . . . . . . . . . . . . . . . . i
中文摘要 . . . . . . . . . . . . . . . . . . . . . . . . . . iii
Abstract . . . . . . . . . . . . . . . . . . . . . . . . . . iv
1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .1
1.1 Thesis Objectives . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.2 Roadmap . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
2 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
2.1 Supervised and Semi-Supervised Learning . . . . . . . . . . . . . . . . .6
2.2 Industrial-Type Datasets . . . . . . . . . . . . . . . . . . . . . . . . 7
2.3 Data Augmentation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
2.4 Mixup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
3 Related Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
3.1 Semi-Supervised Learning . . . . . . . . . . . . . . . . . . . . . . . . 12
3.1.1 Consistency Regularization . . . . . . . . . . . . . . . . . . . . . . 12
3.1.2 Entropy Minimization . . . . . . . . . . . . . . . . . . . . . .. . . 14
3.1.3 Traditional Regularization . . . . . . . . . . . . . . . . . . . . . . 16
3.2 Self-supervised Representation Learning . . . . . . . . . . . . . . . . . 16
3.2.1 Relative Position . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
3.2.2 Image Rotation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
3.2.3 Image Colorization . . . . . . . . . . . . . . . . . . . . . . . . . . 18
3.2.4 AET . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
3.2.5 Auto-encoder . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
3.3 Triplet Loss . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
4 Pseudo-Representation Labeling . . . . . . . . . . . . . . . . . . . . . . 23
4.1 Framework Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
4.2 Batch Pseudo-Labeling . . . . . . . . . . . . . . . . . . . . . . . . . . 24
4.3 Representation Learning Integration . . . . . . . . . . . . . . . . . . . 28
4.4 Feature Space Data Augmentation . . . . . . . . . . . . . . . . . . . 31
4.4.1 VAE on Industrial Dataset . . . . . . . . . . . . . . . . . . . . 32
4.4.2 Feature Space Mixup . . . . . . . . . . . . . . . . . . . . . . . 32
5 Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . 36
5.1 Implementation Details . . . . . . . . . . . . . . . . . . . . . . . . . . 36
5.2 Noise Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
5.3 Utilized labeled Data Size . . . . . . . . . . . . . . . . . . . . . . . . 38
5.4 Circle Relative Position Representation Learning . . . . . . . . . . . . 40
5.5 Representation Learning Improvement . . . . . . . . . . . . . . . . . 42
5.6 Mixup in Different Model Layer . . . . . . . . . . . . . . . . . . . . . 43
5.7 Results Comparison . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
5.8 Labeled Data Amount . . . . . . . . . . . . . . . . . . . . . . . . . . 45
5.9 Batch Size Setting . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
5.10 Comparison of General Dataset . . . . . . . . . . . . . . . . . . . . . 48
6 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
dc.language.isoen
dc.title工業應用之偽標籤特徵半監督式學習zh_TW
dc.titleA Novel Approach, Pseudo-Representation Labeling, for Semi-Supervised Learning on Industrial Applicationsen
dc.typeThesis
dc.date.schoolyear108-2
dc.description.degree碩士
dc.contributor.oralexamcommittee雷欽隆(Chin-Laung Lei),李宏毅(Hung-Yi Lee),王奕翔(I-Hsiang Wang)
dc.subject.keyword半監督學習,自我監督學習,偽標籤,工業應用,資料擴充,zh_TW
dc.subject.keywordSemi-Supervised learning,Self-Supervised Learning,Pseudo Labeling,Industrial Application,Data Augmentation,en
dc.relation.page56
dc.identifier.doi10.6342/NTU202003577
dc.rights.note未授權
dc.date.accepted2020-08-19
dc.contributor.author-college電機資訊學院zh_TW
dc.contributor.author-dept電機工程學研究所zh_TW
顯示於系所單位:電機工程學系

文件中的檔案:
檔案 大小格式 
U0001-1608202013510800.pdf
  未授權公開取用
14.27 MBAdobe PDF
顯示文件簡單紀錄


系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。

社群連結
聯絡資訊
10617臺北市大安區羅斯福路四段1號
No.1 Sec.4, Roosevelt Rd., Taipei, Taiwan, R.O.C. 106
Tel: (02)33662353
Email: ntuetds@ntu.edu.tw
意見箱
相關連結
館藏目錄
國內圖書館整合查詢 MetaCat
臺大學術典藏 NTU Scholars
臺大圖書館數位典藏館
本站聲明
© NTU Library All Rights Reserved