請用此 Handle URI 來引用此文件:
http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/52698完整後設資料紀錄
| DC 欄位 | 值 | 語言 |
|---|---|---|
| dc.contributor.advisor | 王勝德(Sheng-De Wang) | |
| dc.contributor.author | Shang-Lun Tsai | en |
| dc.contributor.author | 蔡尚倫 | zh_TW |
| dc.date.accessioned | 2021-06-15T16:23:46Z | - |
| dc.date.available | 2020-08-25 | |
| dc.date.copyright | 2020-08-25 | |
| dc.date.issued | 2020 | |
| dc.date.submitted | 2020-08-06 | |
| dc.identifier.citation | “Github/PlotNeuralNet ,” https://github.com/HarisIqbal88/PlotNeuralNet. “Types of frequency distribution,” https://getnave.com/blog/frequency-distribution-types/. Y. Cui, M. Jia, T.-Y. Lin, Y. Song, and S. Belongie, “Class-balanced loss based on effective number of samples,” in IEEE Conference on Computer Vision and Pattern Recognition, 2019. A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in Neural Information Processing Systems, 2012. K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in International Conference on Learning Representations, 2015. C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich, “Going deeper with convolutions,” in IEEE Conference on Computer Vision and Pattern Recognition, 2015. “CIFAR dataset,” https://www.cs.toronto.edu/ kriz/cifar.html. “ImageNet dataset,” http://www.image-net.org/. J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, “Imagenet: A largescale hierarchical image database,” in IEEE Conference on Computer Vision and Pattern Recognition, 2009. C. Wah, S. Branson, P. Welinder, P. Perona, and S. Belongie, “The caltech-ucsd birds-200-2011 dataset,” 2011. G. V. Horn, O. M. Aodha, Y. Song, Y. Cui, C. Sun, A. Shepard, H. Adam, P. Perona, and S. Belongie, “The inaturalist species classification and detection dataset,” in IEEE Conference on Computer Vision and Pattern Recognition, 2018. “The Cancer Imaging Archive,” https://www.cancerimagingarchive.net/. “Quora Insincere Questions Classification,” https://www.kaggle.com/c/quora-insincere-questions-classification/overview. “Diabetic retinopathy detection,” https://www.kaggle.com/c/diabetic-retinopathy-detection/overview. M. Buda, A. Maki, and M. A. Mazurowski, “A systematic study of the class imbalance problem in convolutional neural networks,” oct 2018, arXiv:1710.05381. S. M., A. Elrahman, and A. Abraham, “A Review of Class Imbalance Problem,” In Journal of Network and Innovative Computing, 2013. X. G. et al., “On the class imbalance problem,” in In 4th IEEE International Conference on Natural Computation, 2008, pp. 192–201. N. V. Chawla, N. Japkowicz, and A. R. Kołcz, “ Editorial: special issue on learning from imbalanced data sets,” In ACM SIGKDD Explorations Newsletter, 2004. N. V. Chawla, K. W. Bowyer, L. O. Hall, and W. P. Kegelmeyer, “SMOTE: Synthetic Minority Over-sampling Technique,” Journal of Artificial Intelligence Research, vol. 16, pp. 321–357, 2002. C. Drummond and R. C. Holte, “C4.5, Class Imbalance, and Cost Sensitivity: Why Under-Sampling beats Over-Sampling,” in Workshop on Learning from Imbalanced Datasets II, ICML, 2003. Y.-I. Kang and S. Won, “Weight decision algorithm for oversampling technique on class-imbalanced learning,” in International Conference on Control, Automation and Systems, 2010. C. Huang, Y. Li, C. C. Loy, and X. Tang, “Learning Deep Representation for Imbalanced Classification,” in IEEE Conference on Computer Vision and Pattern Recognition, 2016. S. Bengio, “Sharing representations for long tail computer vision problems,” in International Conference on Machine Learning, 2015. Z. Liu, X. Zhan, Z. Miao, J. Wang, B. Gong, and S. X. Yu, “Large-Scale Long-Tailed Recognition in an Open World,” in IEEE Conference on Computer Vision and Pattern Recognition, 2019. Foss, “Heavy tailed and long tailed,” http://www.springer.com/cda/content/document/cda_downloaddocument/9781461471004-c1.pdf?SGWID=0-0-45-1395304-p175250259, oct 2008. G. A., “Long tail: The lecture,” https://web.stanford.edu/~ashishg/msande235/spr08_09/Lecture07.pdf, dec 2017. S. H. Khan, M. Bennamoun, F. Sohel, and R. Togneri, “Cost sensitive learning of deep feature representations from imbalanced data,” aug 2015, arXiv:1508.03422. M. Bucher, S. Herbin, and F. Jurie, “Hard Negative Mining for Metric Learning Based Zero-Shot Classification,” in In ECCV Workshops, no. 3, 2016. T.-Y. Lin, P. Goyal, R. Girshick, K. He, and P. Dollar, “Focal loss for dense object detection,” in IEEE Conference on Computer Vision and Pattern Recognition, 2018. K. Cao, C. Wei, A. Gaidon, N. Arechiga, and T. Ma, “Learning imbalanced datasets with label-distribution-aware margin loss,” in In Advances in Neural Information Processing Systems, 2019. D. Mahajan, R. Girshick, V. Ramanathan, K. He, M. Paluri, Y. Li, A. Bharambe, and van der Maaten L., “Exploring the limits of weakly supervised pretraining,” in European Conference on Computer Vision, 2018. “Wiki/Entropy (information theory),” https://en.wikipedia.org/wiki/Entropy. “Reproduce issue example 1,” https://github.com/richardaecn/class-balanced-loss/issues/5. “Reproduce issue example 2,” https://github.com/kaidic/LDAM-DRW/issues/4. Y. Geifman and R. El-Yaniv, “Deep active learning over the long tail,” nov 2017, arXiv:1711.00941. L. Shen, Z. Lin, and Q. Huang, “Relay backpropagation for effective learning of deep convolutional neural networks,” in European Conference on Computer Vision, 2016. Y. Zou, Z. Yu, B. V. Kumar, and J. Wang, “Unsupervised domain adaptation for semantic segmentation via class-balanced self-training,” in European Conference on Computer Vision, 2018. M. Ren, W. Zeng, B. Yang, and R. Urtasun, “Learning to reweight examples for robust deep learning,” in International Conference on Machine Learning, 2018. Q. Dong, S. Gong, and X. Zhu, “Imbalanced deep learning by minority class incremental rectification,” In IEEE TPAMI, 2018. M. Hayat, S. Khan, W. Zamir, J. Shen, and L. Shao, “Maxmargin class imbalanced learning with gaussian affinity,” in IEEE International Conference on Computer Vision, 2019. J. Shu, Q. Xie, L. Yi, Q. Zhao, S. Zhou, Z. Xu, and D. Meng, “Meta-weight-net:Learning an explicit mapping for sample weighting,” sep 2019, arXiv:1902.07379. Y.Wang,W. Gan,W.Wu, and J. Yan, “Dynamic curriculum learning for imbalanced data classification,” in IEEE International Conference on Computer Vision, 2019. L. E., C. Q., and Q. X., “Deep reinforcement learning for imbalanced classification,” sat 2019, arXiv:1901.01379. R. S. Pereira and F. Porto, “Deep Learning Application for Plant Classification on Unbalanced Training Set,” in In Brazilian e-Science Workshop (BreSci), 2019. A. Ali-Gombe, E. Elyan, and C. Jayne, “Multiple Fake Classes GAN for Data Augmentation in Face Image Dataset,” in International Joint Conference on Neural Networks (IJCNN), 2019, pp. 1–8. L. Huang, K. C. Lin, and Y. Tseng, “Resolving intra-class imbalance for gan-based image augmentation,” in IEEE International Conference on Multimedia and Expo (ICME), 2019, pp. 970–975. G. Mariani, F. Scheidegger, R. Istrate, C. Bekas, and C. Malossi, “BAGAN: Data augmentation with balancing GAN,” jun 2018, arXiv:1803.09655. I. J. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative Adversarial Networks,” in In Neural Information Processing Systems Conference, 2014. A.-G. Adamu, E. Eyad, S. Yann, and J. Chrisina, “Few-shot classifier gan,” in International Joint Conference on Neural Networks (IJCNN), 2018. A. Odena, C. Olah, and J. Shlens, “Conditional image synthesis with auxiliary classifier gans,” dec 2016, arXiv:1610.09585. V. R. Konda and J. N. Tsitsiklis, “Actor-critic algorithms,” in Advances in Neural Information Processing Systems, 2000. M. Bucher, S. Herbin, and F. Jurie, “Improving Semantic Embedding Consistency by Metric Learning for Zero-Shot Classification,” in European Conference on Computer Vision, 2016. S. Janson, “Random coverings in several dimensions. Acta Mathematica,” 1986. “Wiki/Stochastic gradient descent,” https://en.wikipedia.org/wiki/Stochastic_gradient_descent. D. Kingma and J. Ba, “Adam: A method for stochastic optimization,” in International Conference on Learning Representations, 2015. “Wiki/Information content,” https://en.wikipedia.org/wiki/Information_content. J. Yosinski, J. Clune, Y. Bengio, and H. Lipson, “ How transferable are features in deep neural networks?” in Advances in Neural Information Processing Systems, 2014. P. Goyal, P. Doll´ar, R. Girshick, P. Noordhuis, L. Wesolowski, A. Kyrola, A. Tulloch, Y. Jia, and K. He, “Accurate, large minibatch sgd: training imagenet in 1 hour,” jun 2017, arXiv:1706.02677. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in IEEE Conference on Computer Vision and Pattern Recognition, 2016. “Pytorch,” https://pytorch.org/. | |
| dc.identifier.uri | http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/52698 | - |
| dc.description.abstract | 數據不平衡是指訓練數據集中類的傾斜分佈。深度學習演算法訓練類別數據量不平衡資料集時,往往在少數類別上表現較差的預測結果。不幸的是真實世界應用中比如異常偵測,往往存在不同類別間訓練數據差距繁多之疑慮。本論文中,我們針對上述問題提出兩種改善方法分別應用於訓練階段與預測階段。訓練階段時,我們提出類別資訊本體相關之重新加權損失函數,藉由信息增益重新定義各類別之相對權重,使訓練模型將注意力放在少數類別上。預測階段時,我們針對先前的研究與實驗觀察,發現深度學習訓練不平衡資料集所產生之缺陷,藉此延伸解決方案及猜想,提出基於特徵比對之重排序預測,運用先前訓練之特徵抽取器,擷取各類別之特徵範本作為輔助集,藉以比對測試資料與其餘弦相似度,綜合上述輔助預測與原有模型預測結果,採取多數決結果作為最終重排序之預測。實驗上我們使用殘差網路架構作為訓練模型,透過long-tailed CIFAR-10與long-tailed CIFAR-100來檢驗我們提出之方法。實驗結果表明在不平衡數據集上,我們提出的方法顯著提升Top-1準確度。 | zh_TW |
| dc.description.abstract | The skewed distribution of classes in a training dataset refers to as data imbalance. Deep learning algorithms usually suffer poor performance in rare or minor classes when facing the imbalanced datasets. Unfortunately, real world applications typically have the data imbalance issue, such as anomaly detection. In this thesis, we propose two improved methods for the training and the testing stages to solve the data imbalance problem. For the training process, we introduce a Self-information Relevant Re-weighting Loss Function, a theoretical method that redefines the relative weighting of each class through information gained and makes the training model pay attention to minor classes. For the prediction process, based on the previous research and experimental observations, we found the deficiencies generated by deep learning algorithms when training imbalanced datasets. In our solution, the re-ranking method bases on robust feature matching and the voting mechanism. The feature extractor extracts the feature paradigms from each class as a support set, and each type of feature paradigm is used to match the test data with its cosine similarity. The prediction results generated by the support set compare with the original model prediction by majority voting to obtain the final re-ranking prediction. In the experiments, we use the residual network architecture as our training model, and we validate our proposed method through long-tailed CIFAR-10 and long-tailed CIFAR-100. Experimental results show that our method could be able to achieve significant improvement of Top-1 accuracy on imbalanced datasets. | en |
| dc.description.provenance | Made available in DSpace on 2021-06-15T16:23:46Z (GMT). No. of bitstreams: 1 U0001-0608202012394400.pdf: 6512206 bytes, checksum: d9e1cdfc1dce2f5bef5174ee7227149e (MD5) Previous issue date: 2020 | en |
| dc.description.tableofcontents | 口試委員會審定書 i 誌謝 ii 摘要 iii Abstract iv 1 Introduction 1 2 Related Work 6 2.1 Re-sampling Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 2.2 Re-weighting Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 3 Approach 11 3.1 Self-information Relevant Re-weighting Method . . . . . . . . . . . . . 11 3.2 Re-ranking Prediction Based on Feature Matching . . . . . . . . . . . . . 15 4 Experiment 20 4.1 Experiment Setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20 4.2 Experiment Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21 4.3 Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 5 Conclusion Remarks 32 5.1 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32 5.2 Future Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 Bibliography 34 | |
| dc.language.iso | en | |
| dc.subject | 不平衡分類 | zh_TW |
| dc.subject | 重新加權損失函數 | zh_TW |
| dc.subject | 信息理論 | zh_TW |
| dc.subject | 特徵比對 | zh_TW |
| dc.subject | 重排序預測 | zh_TW |
| dc.subject | 深度學習 | zh_TW |
| dc.subject | feature matching | en |
| dc.subject | re-weighting loss function | en |
| dc.subject | information theory | en |
| dc.subject | imbalanced classification | en |
| dc.subject | re-ranking prediction | en |
| dc.subject | deep learning | en |
| dc.title | 基於重新加權損失函數與預測重排序解決深度學習類別訓練資料失衡問題 | zh_TW |
| dc.title | Re-weighting Loss Functions and Re-ranking Predictions for the Class Imbalance Problem in Deep Learning | en |
| dc.type | Thesis | |
| dc.date.schoolyear | 108-2 | |
| dc.description.degree | 碩士 | |
| dc.contributor.oralexamcommittee | 雷欽隆(Chin-Laung Lei),林宗男(Tsung-Nan Lin) | |
| dc.subject.keyword | 不平衡分類,重新加權損失函數,信息理論,特徵比對,重排序預測,深度學習, | zh_TW |
| dc.subject.keyword | imbalanced classification,re-weighting loss function,information theory,feature matching,re-ranking prediction,deep learning, | en |
| dc.relation.page | 38 | |
| dc.identifier.doi | 10.6342/NTU202002524 | |
| dc.rights.note | 有償授權 | |
| dc.date.accepted | 2020-08-06 | |
| dc.contributor.author-college | 電機資訊學院 | zh_TW |
| dc.contributor.author-dept | 電機工程學研究所 | zh_TW |
| 顯示於系所單位: | 電機工程學系 | |
文件中的檔案:
| 檔案 | 大小 | 格式 | |
|---|---|---|---|
| U0001-0608202012394400.pdf 未授權公開取用 | 6.36 MB | Adobe PDF |
系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。
