請用此 Handle URI 來引用此文件:
http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/87195完整後設資料紀錄
| DC 欄位 | 值 | 語言 |
|---|---|---|
| dc.contributor.advisor | 雷欽隆 | zh_TW |
| dc.contributor.advisor | Chin-Laung Lei | en |
| dc.contributor.author | 林家慶 | zh_TW |
| dc.contributor.author | Chia-Ching Lin | en |
| dc.date.accessioned | 2023-05-18T16:18:12Z | - |
| dc.date.available | 2023-11-09 | - |
| dc.date.copyright | 2023-05-10 | - |
| dc.date.issued | 2023 | - |
| dc.date.submitted | 2023-02-09 | - |
| dc.identifier.citation | [1] Bharath Hariharan and Ross Girshick. Low-shot visual recognition by shrinking and hallucinating features. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), pages 3018–3027, 2017.
[2] Yu-Xiong Wang, Ross Girshick, Martial Hebert, and Bharath Hariharan. Low-shot learning from imaginary data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 7278–7286, 2018. [3] Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556, 2014. [4] Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew Rabinovich. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 1–9, 2015. [5] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 770–778, 2016. [6] Ross Girshick, Jeff Donahue, Trevor Darrell, and Jitendra Malik. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 580–587, 2014. [7] Ross Girshick. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), pages 1440–1448, 2015. [8] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster r-cnn: Towards real-time object detection with region proposal networks. In Advances in Neural Information Processing Systems 28 (NIPS 2015), pages 91–99, 2015. [9] Joseph Redmon, Santosh Divvala, Ross Girshick, and Ali Farhadi. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 779–788, 2016. [10] Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 3431–3440, 2015. [11] Liang-Chieh Chen, George Papandreou, Florian Schroff, and Hartwig Adam. Rethinking atrous convolution for semantic image segmentation. arXiv:1706.05587, 2017. [12] Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), pages 2961–2969, 2017. [13] Gary Marcus. Deep learning: A critical appraisal. arXiv:1801.00631, 2018. [14] Amina Adadi. A survey on data‐efficient algorithms in big data era. Journal of Big Data, 8, 2021. [15] Yaqing Wang, Quanming Yao, James T. Kwok, and Lionel M. Ni. Generalizing from a few examples: A survey on few-shot learning. ACM Computing Surveys, 53, 2020. [16] Wen Li, Limin Wang, Wei Li, Eirikur Agustsson, and Luc Van Gool. Webvision database: Visual learning and understanding from web data. arXiv:1708.02862, 2017. [17] Dhruv Mahajan, Ross B. Girshick, Vignesh Ramanathan, Kaiming He, Manohar Paluri, Yixuan Li, Ashwin Bharambe, and Laurens van der Maaten. Exploring the limits of weakly supervised pretraining. In European Conference on Computer Vision (ECCV), 2018. [18] Chiyuan Zhang, Samy Bengio, Moritz Hardt, Benjamin Recht, and Oriol Vinyals. Understanding deep learning requires rethinking generalization. In International Conference on Learning Representations (ICLR), 2017. [19] Devansh Arpit, Stanisław Jastrzębski, Nicolas Ballas, David Krueger, Emmanuel Bengio, Maxinder S. Kanwal, Tegan Maharaj, Asja Fischer, Aaron Courville, Yoshua Bengio, and Simon Lacoste-Julien. A closer look at memorization in deep networks. In Proceedings of the 34th International Conference on Machine Learning (ICML), pages 233–242, 2017. [20] Andreas Eitel, Jost Tobias Springenberg, Luciano Spinello, Martin Riedmiller, and Wolfram Burgard. Multimodal deep learning for robust rgb-d object recognition. In 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 681–687, 2015. [21] Nima Tajbakhsh, Laura Jeyaseelan, Qian Li, Jeffrey N. Chiang, Zhihao Wu, and Xiaowei Ding. Embracing imperfect datasets: A review of deep learning solutions for medical image segmentation. Medical Image Analysis, 63, 2020. [22] Hwanjun Song, Minseok Kim, Dongmin Park, Yooju Shin, and Jae-Gil Lee. Learning from noisy labels with deep neural networks: A survey. IEEE Transactions on Neural Networks and Learning Systems, pages 1–19, 2022. [23] Timothy Hospedales, Antreas Antoniou, Paul Micaelli, and Amos Storkey. Metalearning in neural networks: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(9):5149–5169, 2022. [24] Longlong Jing and Yingli Tian. Self-supervised visual feature learning with deep neural networks: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 43(11):4037–4058, 2021. [25] Oriol Vinyals, Charles Blundell, Timothy Lillicrap, koray kavukcuoglu, and Daan Wierstra. Matching networks for one shot learning. In Advances In Neural Information Processing Systems 29 (NIPS 2016), pages 3630–3638, 2016. [26] Sebastian Thrun and Lorien Y. Pratt. Learning To Learn. Kluwer Academic Publishers, Boston, MA, 1998. [27] Chelsea Finn, Pieter Abbeel, and Sergey Levine. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the 34th International Conference on Machine Learning (ICML), pages 1126–1135, 2017. [28] Hang Qi, Matthew Brown, and David G. Lowe. Low-shot learning with imprinted weights. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 5822–5830, 2018. [29] Spyros Gidaris and Nikos Komodakis. Dynamic few-shot visual learning without forgetting. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 4367–4375, 2018. [30] Wei-Yu Chen, Yen-Cheng Liu, Zsolt Kira, Yu-Chiang Frank Wang, and Jia-Bin Huang. A closer look at few-shot classification. In International Conference on Learning Representations (ICLR), 2019. [31] Yonglong Tian, Yue Wang, Dilip Krishnan, Joshua B. Tenenbaum, and Phillip Isola. Rethinking few-shot image classification: a good embedding is all you need? In European Conference on Computer Vision (ECCV), 2020. [32] Jake Snell, Kevin Swersky, and Richard Zemel. Prototypical networks for few-shot learning. In Advances in Neural Information Processing Systems 30 (NIPS 2017), pages 4077–4087, 2017. [33] Gregory Koch, Richard Zemel, and Ruslan Salakhutdinov. Siamese neural networks for one-shot image recognition. In Proceedings of the 32nd International Conference on Machine Learning (ICML) Workshops, 2015. [34] Flood Sung, Yongxin Yang, Li Zhang, Tao Xiang, Philip H.S. Torr, and Timothy M. Hospedales. Learning to compare: Relation network for few-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 1199–1208, 2018. [35] Sachin Ravi and Hugo Larochelle. Optimization as a model for few-shot learning. In International Conference on Learning Representations (ICLR), 2017. [36] Adam Santoro, Sergey Bartunov, Matthew Botvinick, Daan Wierstra, and Timothy Lillicrap. Meta-learning with memory-augmented neural networks. In Proceedings of The 33rd International Conference on Machine Learning (ICML), pages 1842– 1850, 2016. [37] Eleni Triantafillou, Tyler Zhu, Vincent Dumoulin, Pascal Lamblin, Utku Evci, Kelvin Xu, Ross Goroshin, Carles Gelada, Kevin Swersky, Pierre-Antoine Manzagol, and Hugo Larochelle. Meta-dataset: A dataset of datasets for learning to learn from few examples. In International Conference on Learning Representations (ICLR), 2020. [38] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 248–255, 2009. [39] Brenden M. Lake, Ruslan Salakhutdinov, and Joshua B. Tenenbaum. Humanlevel concept learning through probabilistic program induction. Science, 350(6266):1332–1338, 2015. [40] Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, Yoshua Bengio, Ian J Goodfellow, and Jean Pouget-Abadie. Generative adversarial nets. In Advances in Neural Information Processing Systems 27 (NIPS 2014), pages 2672–2680, 2014. [41] Mehdi Mirza and Simon Osindero. Conditional generative adversarial nets. arXiv:1411.1784, 2014. [42] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial networks. Communications of the ACM, 63(11):139–144, 2020. [43] Kai Li, Yulun Zhang, Kunpeng Li, and Yun Fu. Adversarial feature hallucination networks for few-shot learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 13470–13479, 2020. [44] Zitian Chen, Yanwei Fu, Yinda Zhang, Yu-Gang Jiang, Xiangyang Xue, and Leonid Sigal. Multi-level semantic feature augmentation for one-shot learning. IEEE Transactions on Image Processing, 28(9):4594–4605, 2019. [45] Eli Schwartz, Leonid Karlinsky, Joseph Shtok, Sivan Harary, Mattias Marder, Abhishek Kumar, Rogerio Feris, Raja Giryes, and Alex Bronstein. Delta-encoder: an effective sample synthesis method for few-shot object recognition. In Advances in Neural Information Processing Systems 31 (NeurIPS 2018), pages 2845–2855, 2018. [46] Mengting Chen, Yuxin Fang, Xinggang Wang, Heng Luo, Yifeng Geng, Xinyu Zhang, Chang Huang, Wenyu Liu, and Bo Wang. Diversity transfer network for few-shot learning. In Proceedings of the AAAI Conference on Artificial Intelligence, pages 10559–10566, 2020. [47] Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A. Efros. Unpaired image-toimage translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), pages 2223–2232, 2017. [48] Jun-Yan Zhu, Richard Zhang, Deepak Pathak, Trevor Darrell, Alexei A Efros, Oliver Wang, and Eli Shechtman. Toward multimodal image-to-image translation. In Advances in Neural Information Processing Systems 30 (NIPS 2017), pages 465–476, 2017. [49] Hang Gao, Zheng Shou, Alireza Zareian, Hanwang Zhang, and Shih-Fu Chang. Low-shot learning via covariance-preserving adversarial augmentation networks. In Advances in Neural Information Processing Systems 31 (NeurIPS 2018), pages 975–985, 2018. [50] Zitian Chen, Yanwei Fu, Yu-Xiong Wang, Lin Ma, Wei Liu, and Martial Hebert. Image deformation meta-networks for one-shot learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 8680–8689, 2019. [51] Satoshi Tsutsui, Yanwei Fu, and David Crandall. Meta-reinforced synthetic data for one-shot fine-grained visual recognition. In Advances in Neural Information Processing Systems 32 (NeurIPS 2019), pages 3063–3072, 2019. [52] Hongguang Zhang, Jing Zhang, and Piotr Koniusz. Few-shot learning via saliencyguided hallucination of samples. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 2770–2779, 2019. [53] Benoit Frenay and Michel Verleysen. Classification in the presence of label noise: A survey. IEEE Transactions on Neural Networks and Learning Systems, 25(5):845–869, 2014. [54] Aritra Ghosh, Himanshu Kumar, and P.S. Sastry. Robust loss functions under label noise for deep neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, 2017. [55] Zhilu Zhang and Mert Sabuncu. Generalized cross entropy loss for training deep neural networks with noisy labels. In Advances in Neural Information Processing Systems 31 (NeurIPS 2018), pages 8778–8788, 2018. [56] Yisen Wang, Xingjun Ma, Zaiyi Chen, Yuan Luo, Jinfeng Yi, and James Bailey. Symmetric cross entropy for robust learning with noisy labels. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 322–330, 2019. [57] Yueming Lyu and Ivor W. Tsang. Curriculum loss: Robust learning and generalization against label corruption. In International Conference on Learning Representations (ICLR), 2020. [58] Jacob Goldberger and Ehud Ben-Reuven. Training deep neural-networks using a noise adaptation layer. In International Conference on Learning Representations (ICLR), 2017. [59] Giorgio Patrini, Alessandro Rozza, Aditya Krishna Menon, Richard Nock, and Lizhen Qu. Making deep neural networks robust to label noise: A loss correction approach. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 1944–1952, 2017. [60] Dan Hendrycks, Mantas Mazeika, Duncan Wilson, and Kevin Gimpel. Using trusted data to train deep networks on labels corrupted by severe noise. In Advances in Neural Information Processing Systems 31 (NeurIPS 2018), pages 10456–10465, 2018. [61] Xiaobo Xia, Tongliang Liu, Nannan Wang, Bo Han, Chen Gong, Gang Niu, and Masashi Sugiyama. Are anchor points really indispensable in label-noise learning? In Advances in Neural Information Processing Systems 32 (NeurIPS 2019), pages 6838–6849, 2019. [62] Zhen Wang, Guosheng Hu, and Qinghua Hu. Training noise-robust deep neural networks via meta-learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 4524–4533, 2020. [63] Yu Yao, Tongliang Liu, Bo Han, Mingming Gong, Jiankang Deng, Gang Niu, and Masashi Sugiyama. Dual t: Reducing estimation error for transition matrix in labelnoise learning. In Advances in Neural Information Processing Systems 33 (NeurIPS 2020), 2020. [64] Dan Hendrycks, Mantas Mazeika, Saurav Kadavath, and Dawn Song. Using selfsupervised learning can improve model robustness and uncertainty. In Advances in Neural Information Processing Systems 32 (NeurIPS 2019), pages 15663–15674, 2019. [65] Aritra Ghosh and Andrew Lan. Contrastive learning improves model robustness under label noise. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, pages 2703–2708, 2021. [66] Yazhou Yao, Zeren Sun, Chuanyi Zhang, Fumin Shen, Qi Wu, Jian Zhang, and Zhenmin Tang. Jo-src: A contrastive approach for combating noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 5192–5201, 2021. [67] Diego Ortego, Eric Arazo, Paul Albert, Noel E. O’Connor, and Kevin McGuinness. Multi-objective interpolation training for robustness to label noise. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 6606–6615, 2021. [68] Spyros Gidaris, Praveer Singh, and Nikos Komodakis. Unsupervised representation learning by predicting image rotations. In International Conference on Learning Representations (ICLR), 2018. [69] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In Proceedings of the 37th International Conference on Machine Learning (ICML), pages 1597–1607, 2020. [70] Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, Bilal Piot, koray kavukcuoglu, Remi Munos, and Michal Valko. Bootstrap your own latent - a new approach to self-supervised learning. In Advances in Neural Information Processing Systems 33 (NeurIPS 2020), pages 21271–21284, 2020. [71] Xinlei Chen and Kaiming He. Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 15750–15758, 2021. [72] Junnan Li, Yongkang Wong, Qi Zhao, and Mohan S. Kankanhalli. Learning to learn from noisy labeled data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 5051–5059, 2019. [73] Eran Malach and Shai Shalev-Shwartz. Decoupling ”when to update” from ”how to update”. In Advances in Neural Information Processing Systems 30 (NIPS 2017), pages 960–970, 2017. [74] Lu Jiang, Zhengyuan Zhou, Thomas Leung, Li-Jia Li, and Li Fei-Fei. MentorNet: Learning data-driven curriculum for very deep neural networks on corrupted labels. In Proceedings of the 35th International Conference on Machine Learning (ICML), pages 2304–2313, 2018. [75] Bo Han, Quanming Yao, Xingrui Yu, Gang Niu, Miao Xu, Weihua Hu, Ivor Tsang, and Masashi Sugiyama. Co-teaching: Robust training of deep neural networks with extremely noisy labels. In Advances in Neural Information Processing Systems 31 (NeurIPS 2018), pages 8527–8537, 2018. [76] Xingrui Yu, Bo Han, Jiangchao Yao, Gang Niu, Ivor Tsang, and Masashi Sugiyama. How does disagreement help generalization against label corruption? In Proceedings of the 36th International Conference on Machine Learning (ICML), pages 7164–7173, 2019. [77] Hongxin Wei, Lei Feng, Xiangyu Chen, and Bo An. Combating noisy labels by agreement: A joint training method with co-regularization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 13726–13735, 2020. [78] Yanyao Shen and Sujay Sanghavi. Learning with bad training data via iterative trimmed loss minimization. In Proceedings of the 36th International Conference on Machine Learning (ICML), pages 5739–5748, 2019. [79] Pengfei Chen, Ben Ben Liao, Guangyong Chen, and Shengyu Zhang. Understanding and utilizing deep neural networks trained with noisy labels. In Proceedings of the 36th International Conference on Machine Learning (ICML), pages 1062–1070, 2019. [80] Jinchi Huang, Lie Qu, Rongfei Jia, and Binqiang Zhao. O2u-net: A simple noisy label detection approach for deep neural networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 3326–3334, 2019. [81] Mengye Ren, Wenyuan Zeng, Bin Yang, and Raquel Urtasun. Learning to reweight examples for robust deep learning. In Proceedings of the 35th International Conference on Machine Learning (ICML), pages 4334–4343, 2018. [82] Jun Shu, Qi Xie, Lixuan Yi, Qian Zhao, Sanping Zhou, Zongben Xu, and Deyu Meng. Meta-weight-net: Learning an explicit mapping for sample weighting. In Advances in Neural Information Processing Systems 32 (NeurIPS 2019), pages 1919–1930, 2019. [83] Jiangfan Han, Ping Luo, and Xiaogang Wang. Deep self-learning from noisy labels. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 5138–5147, 2019. [84] Daiki Tanaka, Daiki Ikami, Toshihiko Yamasaki, and Kiyoharu Aizawa. Joint optimization framework for learning with noisy labels. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 5552– 5560, 2018. [85] Kun Yi and Jianxin Wu. Probabilistic end-to-end noise correction for learning with noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 7017–7025, 2019. [86] Yikai Zhang, Songzhu Zheng, Pengxiang Wu, Mayank Goswami, and Chao Chen. Learning with feature-dependent label noise: A progressive approach. In International Conference on Learning Representations (ICLR), 2021. [87] Zizhao Zhang, Han Zhang, Sercan O. Arik, Honglak Lee, and Tomas Pfister. Distilling effective supervision from severe label noise. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 9294–9303, 2020. [88] Junnan Li, Richard Socher, and Steven C.H. Hoi. Dividemix: Learning with noisy labels as semi-supervised learning. In International Conference on Learning Representations (ICLR), 2020. [89] Kento Nishi, Yi Ding, Alex Rich, and Tobias Hollerer. Augmentation strategies for learning with noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 8022–8031, 2021. [90] Nazmul Karim, Mamshad Nayeem Rizve, Nazanin Rahnavard, Ajmal Mian, and Mubarak Shah. Unicon: Combating label noise through uniform selection and contrastive learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 9676–9686, 2022. [91] Haobo Wang, Ruixuan Xiao, Yiwen Dong, Lei Feng, and Junbo Zhao. Promix: Combating label noise via maximizing clean sample utility. arXiv:2007.08199v6, 2021. [92] Chia-Ching Lin, Yu-Chiang Frank Wang, Chin-Laung Lei, and Kuan-Ta Chen. Semantics-guided data hallucination for few-shot visual classification. In 2019 IEEE International Conference on Image Processing (ICIP), pages 3302–3306, 2019. [93] Yongqin Xian, Christoph H. Lampert, Bernt Schiele, and Zeynep Akata. Zeroshot learning—a comprehensive evaluation of the good, the bad and the ugly. IEEE Transactions on Pattern Analysis and Machine Intelligence, 41(9):2251–2265, 2019. [94] Frederik Pahde, Patrick Jähnichen, Tassilo Klein, and Moin Nabi. Cross-modal hallucination for few-shot fine-grained recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2018. [95] Max Welling Diederik P Kingma. Auto-encoding variational bayes. In International Conference on Learning Representations (ICLR), 2014. [96] Alex Krizhevsky. Learning multiple layers of features from tiny images. Technical report, Toronto, ON, Canada, 2009. [97] Matthew Honnibal and Mark Johnson. An improved non-monotonic transition system for dependency parsing. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1373–1378, 2015. [98] Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. Journal of Machine Learning Research, 9(11):2579––2605, 2008. [99] Joshua B Tenenbaum, Vin de Silva, and John C Langford. A global geometric framework for nonlinear dimensionality reduction. Science, 290(5500):2319– 2323, 2000. [100] Catherine Wah, Steve Branson, Peter Welinder, Pietro Perona, and Serge Belongie. The caltech-ucsd birds-200-2011 dataset. Technical report, California Institute of Technology, 2011. [101] Li Tang, Yue Wang, Qianhui Luo, Xiaqing Ding, and Rong Xiong. Adversarial feature disentanglement for place recognition across changing appearance. In 2020 IEEE International Conference on Robotics and Automation (ICRA), pages 1301–1307, 2020. [102] Dominik Lorenz, Leonard Bereska, Timo Milbich, and Bjorn Ommer. Unsupervised part-based disentangling of object shape and appearance. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 10955–10964, 2019. [103] Bin Cheng, Tao Dai, Bin Chen, Shutao Xia, and Xiu Li. Efficient face manipulation via deep feature disentanglement and reintegration net. In ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1990–1994, 2021. [104] Liu Liu, Jiangtong Li, Li Niu, Ruicong Xu, and Liqing Zhang. Activity imageto- video retrieval by disentangling appearance and motion. In Proceedings of the AAAI Conference on Artificial Intelligence, pages 2145–2153, 2021. [105] Chia-Ching Lin, Hsin-Li Chu, Yu-Chiang Frank Wang, and Chin-Laung Lei. Joint feature disentanglement and hallucination for few-shot image classification. IEEE Transactions on Image Processing, 30:9245–9258, 2021. [106] Geoffrey Hinton and Ruslan Salakhutdinov. Reducing the dimensionality of data with neural networks. Science, 313(5786):504–507, 2006. [107] Hsin-Ying Lee, Hung-Yu Tseng, Jia-Bin Huang, Maneesh Singh, and Ming-Hsuan Yang. Diverse image-to-image translation via disentangled representations. In Proceedings of the European Conference on Computer Vision (ECCV), pages 35–51, 2018. [108] Ming-Yu Liu, Xun Huang, Arun Mallya, Tero Karras, Timo Aila, Jaakko Lehtinen, and Jan Kautz. Few-shot unsupervised image-to-image translation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 10551–10560, 2019. [109] Nathan Hilliard, Lawrence Phillips, Scott Howland, Artëm Yankov, Courtney D. Corley, and Nathan O. Hodas. Few-shot learning with metric-agnostic conditional embeddings. arXiv:1802.04376, 2018. [110] Maria-Elena Nilsback and Andrew Zisserman. Automated flower classification over a large number of classes. In 2008 Sixth Indian Conference on Computer Vision, Graphics & Image Processing (ICVGIP), pages 722–729, 2008. [111] Connor Shorten and Taghi M. Khoshgoftaar. A survey on image data augmentation for deep learning. Journal of Big Data, 6(1):1–48, 2019. [112] Anders Krogh and John Hertz. A simple weight decay can improve generalization. In Advances in Neural Information Processing Systems 4 (NIPS 1991), pages 950– 957, 1991. [113] Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. Dropout: a simple way to prevent neural networks from overfitting. Journal of Machine Learning Research, 15(1):1929–1958, 2014. [114] Sergey Ioffe and Christian Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the 32nd International Conference on Machine Learning (ICML), pages 448–456, 2015. [115] Sainbayar Sukhbaatar, Joan Bruna, Manohar Paluri, Lubomir Bourdev, and Rob Fergus. Training convolutional networks with noisy labels. In International Conference on Learning Representations (ICLR) Workshops, 2015. [116] Alan Joseph Bekker and Jacob Goldberger. Training deep neural-networks based on unreliable labels. In 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 2682–2686, 2016. [117] Jiaheng Wei, Zhaowei Zhu, Hao Cheng, Tongliang Liu, Gang Niu, and Yang Liu. Learning with noisy labels revisited: A study using real-world human annotations. In International Conference on Learning Representations (ICLR), 2022. [118] Tong Xiao, Tian Xia, Yi Yang, Chang Huang, and Xiaogang Wang. Learning from massive noisy labeled data for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 2691–2699, 2015. [119] Mehdi Noroozi and Paolo Favaro. Unsupervised learning of visual representations by solving jigsaw puzzles. In European Conference on Computer Vision (ECCV), 2016. [120] Richard Zhang, Phillip Isola, and Alexei A Efros. Colorful image colorization. In European Conference on Computer Vision (ECCV), 2016. [121] Jiang Lu, Sheng Jin, Jian Liang, and Changshui Zhang. Robust few-shot learning for user-provided data. IEEE Transactions on Neural Networks and Learning Systems, 32(4):1433–1447, 2021. [122] Kevin J. Liang, Samrudhdhi B. Rangrej, Vladan Petrovic, and Tal Hassner. Fewshot learning with noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 9089–9098, 2022. [123] Hung-Yu Tseng, Hsin-Ying Lee, Jia-Bin Huang, and Ming-Hsuan Yang. Crossdomain few-shot classification via learned feature-wise transformation. In International Conference on Learning Representations (ICLR), 2020. [124] Yunhui Guo, Noel C. Codella, Leonid Karlinsky, James V. Codella, John R. Smith, Kate Saenko, Tajana Rosing, and Rogerio Feris. A broader study of cross-domain few-shot learning. In European Conference on Computer Vision (ECCV), 2020. [125] Xiaobo Xia, Tongliang Liu, Bo Han, Nannan Wang, Mingming Gong, Haifeng Liu, Gang Niu, Dacheng Tao, and Masashi Sugiyama. Part-dependent label noise: Towards instance-dependent label noise. In Advances in Neural Information Processing Systems 33 (NeurIPS 2020), pages 7597–7610, 2020. [126] Pengfei Chen, Junjie Ye, Guangyong Chen, Jingwei Zhao, and Pheng-Ann Heng. Beyond class-conditional assumption: A primary attempt to combat instancedependent label noise. In Proceedings of the AAAI Conference on Artificial Intelligence, pages 11442–11450, 2021. [127] Feng Cen, Xiaoyu Zhao, Wuzhuang Li, and Guanghui Wang. Deep feature augmentation for occluded image classification. Pattern Recognition, 111:107737, 2021. [128] Yiming Li, Yong Jiang, Zhifeng Li, and Shu-Tao Xia. Backdoor learning: A survey. IEEE Transactions on Neural Networks and Learning Systems, pages 1–18, 2022. | - |
| dc.identifier.uri | http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/87195 | - |
| dc.description.abstract | 深度學習近年來於電腦視覺之多項任務中均取得突破性的成果,例如影像辨識、物件偵測、圖像語意分割等等。然而,深度學習模型的強大效能,往往來自於龐大且具有精確標註的訓練資料。受限於時間、金錢,與人力等因素,在現實世界的應用場景當中,研究人員往往必須面對不完美的訓練資料,例如數量有限或是帶有錯誤標籤的樣本。因此,如何在面對此類不完美的訓練資料時,仍然能夠訓練出可靠的深度學習模型,已然成為重要的研究主題。本研究分成兩個部分,第一部分探討有限的訓練資料,討論少樣本學習(few-shot learning);第二部分則是探討訓練資料數量充足,但是帶有錯誤標籤的情形,討論噪聲標籤學習(noisy-label learning)。在少樣本學習的研究中,我們從資料擴增(data augmentation)的角度切入,模仿人類的學習模式:藉由觀察基礎類別(base categories)之大量訓練資料,學習舉一反三的能力,從而在面對新類別(novel categories)之少量學習樣本時,有能力幻想(hallucinate)出更多的新類別樣本。現有的資料幻想技術,並沒有充分利用人類學習新類別時可取得的各種先驗知識(prior knowledge),例如類別的語意資訊,或是基礎類別之大量訓練資料當中所隱含的豐富外觀資訊等等。此研究藉由將上述兩種資訊,藉由元學習(meta learning)的技巧,納入資料幻想模型的學習框架當中,並探討他們如何增進深度學習模型在少樣本學習上的表現。實驗結果顯示,我們的模型在提升少樣本類別之準確率,以及生成資料之可解釋性方面,均有比現有方法較佳的表現。在噪聲標籤學習的研究中,我們將自監督學習(self-supervised learning)的概念延伸至集合的層次,刻意擾亂一個小批次(mini-batch)的訓練資料之標籤,並透過元學習的技巧,鼓勵模型在觀察到該標籤擾亂之後,仍然能夠與另一個未觀察到該標籤擾亂之穩定模型具有相似的表現,從而提高模型對錯誤標籤的容忍度。本研究更進一步將上述之集合層次的自監督學習框架進行延伸,發展出估計整體訓練資料當中不同類別之間的標記錯誤機率之演算法,使模型有機會針對特定標記錯誤進行補償;也發展出一種新的樣本重新加權(sample reweighting)演算法,自動調降標記錯誤之訓練樣本的權重,從而提升模型對於錯誤標記之訓練資料的穩健性。實驗結果顯示,我們提出的集合層次自監督學習 (set-level self-supervised learning, SLSSL) 可以有效提升模型在正確標籤之測試資料集的準確率。 | zh_TW |
| dc.description.abstract | Deep learning approaches have recently achieved remarkable progress in many computer vision tasks, such as visual classification, object detection, and semantic segmentation. However, training a deep neural network typically requires a large amount of accurately-labeled training data, which is time-consuming and labor-intensive to collect. In many real-world applications, researchers often have to build their networks based on imperfect training data, such as datasets with a limited number of samples or noisy labels. As a result, how to effectively learn from such imperfect datasets has become an important research topic. This study can be divided into two parts. In the first part, we focus on learning from limited training data, and discuss few-shot learning (FSL) tasks. In the second part, we focus on learning from noisily-labeled training data, which are assumed to have a sufficient amount of samples with noisy labels, and discuss noisy-label learning (NLL) tasks.
In the study of FSL, we start with data augmentation approaches, in which a data hallucinator is built from base categories with a large number of training samples, and then applied to generate additional training samples for few-shot novel categories. Particularly, we leverage the meta-learning technique to investigate two forms of prior knowledge to aid the construction of the hallucinator: 1) the semantic information among all categories; 2) the abundant appearance information extracted from base categories. Experiments show that, by leveraging the above prior knowledge, the proposed frameworks are able to make hallucinations that not only improve the performances of downstream FSL tasks, but also exhibit more explainable appearances or distributions than previous data hallucination works. In the study of NLL, we extend the idea of self-supervised learning to the set level, and propose the set-level self-supervised learning (SLSSL) technique. Specifically, we first corrupt the labels of a training set (i.e., mini-batch) and then obtain an augmented model by the meta-learning techniques. Next, we design our pretext task by imposing a consistency constraint on such augmented models to encourage the model's robustness against noisy labels. The proposed SLSSL technique can also be utilized to estimate the associated noise transition matrix to counteract the effect of noisy labels during model training, and also allows us to identify clean/noisy training samples via sample reweighting. Such SLSSL-based model training and sample reweighting algorithms can be further combined as an EM-like algorithm to boost the NLL performance. Experiments on synthetic and real-world noisily-labeled data confirm the effectiveness of our framework. | en |
| dc.description.provenance | Submitted by admin ntu (admin@lib.ntu.edu.tw) on 2023-05-18T16:18:12Z No. of bitstreams: 0 | en |
| dc.description.provenance | Made available in DSpace on 2023-05-18T16:18:12Z (GMT). No. of bitstreams: 0 | en |
| dc.description.tableofcontents | 口試委員會審定書 i
誌謝 iii 摘要 v Abstract vii Contents xi List of Figures xv List of Tables xvii Chapter 1 Introduction 1 Chapter 2 Preliminary 7 2.1 Few-Shot Learning 7 2.1.1 FSL Scenarios 7 2.1.2 FSL Approaches 10 2.1.2.1 Parameter-Level Approaches 10 2.1.2.2 Data Hallucination Approaches 11 2.1.2.3 Meta-Learned Hallucinator 15 2.2 Noisy-Label Learning 16 2.2.1 Noise Types 16 2.2.2 NLL Approaches 17 2.2.2.1 Parameter-Level Approaches 18 2.2.2.2 Data-Level Approaches 20 I Learning from Limited Data 21 Chapter 3 Semantics-Guided Hallucination for Few-Shot Learning 23 3.1 Overview 23 3.2 Problem Definition 24 3.3 Semantics-Guided Hallucination 25 3.3.1 Generator as Data Hallucinator 26 3.3.2 Semantics-Guided Hallucination Network 27 3.3.3 Procedures of Training and Inference 28 3.4 Experiments 30 3.4.1 Datasets 30 3.4.2 Implementation Details 32 3.4.3 Results 32 3.4.3.1 CIFAR-100 32 3.4.3.2 Animals with Attributes 34 3.5 Summary 35 Chapter 4 Feature Disentanglement based Hallucination for Few-Shot Learning 37 4.1 Overview 37 4.2 Problem Definition 39 4.3 Appearance-Guided Hallucination 41 4.3.1 Preserving Categorical Information 42 4.3.2 Preserving Appearance Information 45 4.3.3 Optimization Procedure 47 4.3.4 Data Hallucination for Few-shot Learning 48 4.4 Experiments 49 4.4.1 Datasets 49 4.4.2 FSL Settings and Implementation Details 50 4.4.3 Evalutation 55 4.4.4 Ablation Study 59 4.5 Summary 62 II Learning from Noisily-labeled Data 63 Chapter 5 Set-Level Self-Supervised Learning from Noisily-Labeled Data 65 5.1 Overview 65 5.2 Problem Definition 66 5.3 Set-Level Self-Supervised Learning 67 5.3.1 SLSSL Operation 67 5.3.2 SLSSL for Model Training 68 5.3.3 SLSSL for Sample Reweighting 74 5.3.4 SLSSL as an EM-like Algorithm 77 5.4 Experiments 78 5.4.1 Datasets 78 5.4.2 Implementation details 79 5.4.3 Quantitative Results 82 5.4.4 Visualization and Analysis 86 5.4.5 Limitations 89 5.5 Summary 90 Chapter 6 Conclusion and Future Works 91 References 97 | - |
| dc.language.iso | en | - |
| dc.subject | 元學習 | zh_TW |
| dc.subject | 自監督學習 | zh_TW |
| dc.subject | 噪聲標籤學習 | zh_TW |
| dc.subject | 少樣本學習 | zh_TW |
| dc.subject | noisy-label learning | en |
| dc.subject | few-shot learning | en |
| dc.subject | self-supervised learning | en |
| dc.subject | meta learning | en |
| dc.title | 有限或帶噪標籤資料學習於視覺分類 | zh_TW |
| dc.title | Learning from Limited or Noisily-Labeled Data for Visual Classification | en |
| dc.type | Thesis | - |
| dc.date.schoolyear | 111-1 | - |
| dc.description.degree | 博士 | - |
| dc.contributor.coadvisor | 王鈺強 | zh_TW |
| dc.contributor.coadvisor | Yu-Chiang Frank Wang | en |
| dc.contributor.oralexamcommittee | 林彥宇;陳祝嵩;王銘宏;于天立;顏嗣鈞;王勝德 | zh_TW |
| dc.contributor.oralexamcommittee | Yen-Yu Lin;Chu-Song Chen;Ming-Hung Wang;Tian-Li Yu;Hsu-chun Yen;Sheng-De Wang | en |
| dc.subject.keyword | 少樣本學習,噪聲標籤學習,元學習,自監督學習, | zh_TW |
| dc.subject.keyword | few-shot learning,noisy-label learning,meta learning,self-supervised learning, | en |
| dc.relation.page | 115 | - |
| dc.identifier.doi | 10.6342/NTU202300302 | - |
| dc.rights.note | 同意授權(全球公開) | - |
| dc.date.accepted | 2023-02-10 | - |
| dc.contributor.author-college | 電機資訊學院 | - |
| dc.contributor.author-dept | 電機工程學系 | - |
| dc.date.embargo-lift | 2025-03-01 | - |
| 顯示於系所單位: | 電機工程學系 | |
文件中的檔案:
| 檔案 | 大小 | 格式 | |
|---|---|---|---|
| ntu-111-1.pdf | 13.81 MB | Adobe PDF | 檢視/開啟 |
系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。
