Skip navigation

DSpace

機構典藏 DSpace 系統致力於保存各式數位資料(如:文字、圖片、PDF)並使其易於取用。

點此認識 DSpace
DSpace logo
English
中文
  • 瀏覽論文
    • 校院系所
    • 出版年
    • 作者
    • 標題
    • 關鍵字
    • 指導教授
  • 搜尋 TDR
  • 授權 Q&A
    • 我的頁面
    • 接受 E-mail 通知
    • 編輯個人資料
  1. NTU Theses and Dissertations Repository
  2. 電機資訊學院
  3. 生醫電子與資訊學研究所
請用此 Handle URI 來引用此文件: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/21591
完整後設資料紀錄
DC 欄位值語言
dc.contributor.advisor傅楸善
dc.contributor.authorTz-Chia Tsengen
dc.contributor.author曾子家zh_TW
dc.date.accessioned2021-06-08T03:39:00Z-
dc.date.copyright2019-07-26
dc.date.issued2019
dc.date.submitted2019-07-15
dc.identifier.citation[1] A. Agarwal, et al., “Face anti-spoofing using haralick features,” In Biometrics Theory, Applications and Systems (BTAS), 2016 IEEE 8th International Conference, pp. 1-6, IEEE, 2016.
[2] Y. Agarwal, “Create your first Image Recognition Classifier using CNN, Keras and Tensorflow backend,” https://medium.com/nybles/create-your-first-image-recognition-classifier-using-cnn-keras-and-tensorflow-backend-6eaab98d14dd, 2019.
[3] A. Anjos and S. Marcel, “Counter-measures to photo attacks in face recognition: A public database and a baseline,” In IJCB, 2011.
[4] A. Anjos, et al., “Face anti-spoofing: Visual approach” In Handbook of Biometric Anti-Spoofing, pp. 65-82, Springer, 2014.
[5] W. Bao, et al., “A liveness detection method for face recognition based on optical flow field,” In Image Analysis and Signal Processing, 2009, IASP 2009 International Conference, pp. 233-236, IEEE, 2009.
[6] P. Bhandari, “Depth Estimation Techniques,” https://medium.com/@piyush1995bhandari/depth-estimation-techniques-830ffd297245, 2018.
[7] Z. Boulkenafet, et al., “Face anti-spoofing based on color texture analysis,” CoRR, abs/1511.06316, 2015.
[8] P. P. K. Chan, et al., “Face Liveness Detection Using a Flash against 2D Spoofing Attack,” IEEE Transactions on Information Forensics and Security, Volume 13, Issue 2, pp. 521–534, 2018.
[9] I. Chingovska, et al., “On the effectiveness of local binary patterns in face anti-spoofing,” BIOSIG, pp. 1-7, 2012.
[10] Common Objects in Context, http://cocodataset.org/#home, 2019.
[11] E. D. Cubuk, et al., “AutoAugment: Learning Augmentation Policies from Data,” arXiv:1805.09501, 2018.
[12] DeepLearning.Hub, “Learning non-maximum suppression,” https://twitter.com/DLdotHub/status/861966336527872000, 2019.
[13] V. Dumoulin, F. Visin, “A guide to convolution arithmetic for deep learning,” arXiv:1603.07285, 2018.
[14] N. Eineck, et al., “A multi-block-matching approach for stereo,” 2015 IEEE Intelligent Vehicles Symposium (IV), 2015.
[15] W. Forson, “Understanding SSD MultiBox — Real-Time Object Detection In Deep Learning,” https://towardsdatascience.com/understanding-ssd-multibox-real-time-object-detection-in-deep-learning-495ef744fab, 2019.
[16] H. Gao, et al., “Understand Single Shot MultiBox Detector (SSD) and Implement It in Pytorch,” https://medium.com/@smallfishbigsea/understand-ssd-and-implement-your-own-caa3232cd6ad, 2019.
[17] R. GmbH, “Stereo matching,” https://doc.rc-visard.com/latest/en/stereo_matching.html#stereo-matching, 2018.
[18] Google Open Source, “Open Images Dataset,” https://opensource.google.com/projects/open-images-dataset, 2019.
[19] R. A. Hamzah1, and H. Ibrahim, “Literature Survey on Stereo Vision Disparity Map Algorithms,” https://www.hindawi.com/journals/js/2016/8742920/, 2016.
[20] H. Hirschmuller, “Stereo Processing by Semi-Global Matching and Mutual Information,” IEEE transactions on pattern analysis and machine, 2007.
[21] Kaggle , “How to choose CNN Architecture MNIST,” https://www.kaggle.com/cdeotte/how-to-choose-cnn-architecture-mnist, 2019.
[22] R. Khandelwal, “Neural Network -Activation functions ,” https://medium.com/datadriveninvestor/neural-networks-activation-functions-e371202b56ff, 2019.
[23] W. Kim, et al., “Face liveness detection from a single image via diffusion speed model,” IEEE Transactions on Image Processing, 24(8):2456-2465, 2015.
[24] K. Kollreider, et al., “Evaluating liveness by face images and the structure tensor,” In IEEE Workshop on Automatic Identification Advanced Technologies, pp. 75-80, 2005.
[25] N. Lakshminarayana, et al., “A discriminative spatio-temporal mapping of face for liveness detection,” 2017 IEEE International Conference on Identity, Security and Behavior Analysis (ISBA), 2017.
[26] W. Liu, et al., “SSD: Single Shot MultiBox Detector,” arXiv:1512.02325, 2016.
[27] J. Maatta, et al., “Face spoofing detection from single images using micro-texture analysis,” Biometrics (IJCB), 2011, international joint conference on, IEEE, pp. 1-7, 2011.
[28] D. T. Nguyen, et al., “Combining Deep and Handcrafted Image Features for Presentation Attack Detection in Face Recognition Systems Using Visible-Light Camera Sensors,” arXiv:1804.06702, 2018.
[29] OpenCV, “Depth Map from Stereo Images,” https://docs.opencv.org/3.1.0/dd/d53/tutorial_py_depthmap.html, 2019.
[30] OpenCV, “Histogram Equalization,” https://opencv-python-tutroals.readthedocs.io/en/latest/py_tutorials/py_imgproc/py_histograms/py_histogram_equalization/py_histogram_equalization.html, 2019.
[31] P. Pai, “Data Augmentation Techniques in CNN using Tensorflow,” https://medium.com/ymedialabs-innovation/data-augmentation-techniques-in-cnn-using-tensorflow-371ae43d5be9, 2019.
[32] G. Pan, et al., “Eyeblink-based anti-spoofing in face recognition from a generic webcamera,” 2007 IEEE 11th International Conference on Computer Vision, 2007.
[33] PASCAL2 Network, “The PASCAL Visual Object Classes Homepage,” http://host.robots.ox.ac.uk/pascal/VOC/, 2019.
[34] T. F. Pereira, et al., “Face liveness detection using dynamic texture.,” EURASIP JIVP, 2014.
[35] X. Qi, et al., “Comparison of Support Vector Machine and Softmax Classifiers in Computer Vision,” 2017 Second International Conference on Mechanical, Control and Computer Engineering, 2017.
[36] J. Redmon, et al., “You Only Look Once: Unified, Real-Time Object Detection,” arXiv:1506.02640, 2019.
[37] S. Ren, et al., “Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks,” arXiv:1506.01497, 2015.
[38] X. Song, et al., “Discriminative Representation Combinations for Accurate Face Spoofing Detection,” arXiv:1808.08802, 2018.
[39] Stanford University, “CS231n Convolutional Neural Networks for Visual Recognition,” http://cs231n.github.io/neural-networks-1/, 2019.
[40] D. Tang, et al., “Face Flashing: a Secure Liveness Detection Protocol Based on Light Reflections,” arXiv:1801.01949, 2018.
[41] R. Tronci, et al., “Fusion of multiple clues for photo-attack detection in face recognition systems,” In 2011 IEEE International Joint Conference on Biometrics, IJCB 2011, Washington, DC, USA, pp. 1-6, 2011.
[42] S. T. Tsang, “Review: VGGNet — 1st Runner-Up (Image Classification), Winner (Localization) in ILSVRC 2014,” https://medium.com/coinmonks/paper-review-of-vggnet-1st-runner-up-of-ilsvlc-2014-image-classification-d02355543a11, 2018.
[43] S. -H. Tsang, “Review: SSD — Single Shot Detector (Object Detection),” https://towardsdatascience.com/review-ssd-single-shot-detector-object-detection-851a94607d11, 2019.
[44] UFLDL Tutorial , “Convolutional Neural Network,” http://deeplearning.stanford.edu/tutorial/supervised/ConvolutionalNeuralNetwork/, 2019.
[45] J. Li, Y. Wang, et al., “Live face detection based on the analysis of Fourier spectra,” In Biometric Technology for Human Identification, pp. 296-303, 2004.
[46] Wikipedia, ”Adaptive histogram equalization,” https://en.wikipedia.org/wiki/Adaptive_histogram_equalization, 2019.
[47] Wikipedia, “Field-programmable gate array,” https://en.wikipedia.org/wiki/Field-programmable_gate_array, 2019.
[48] Wikipedia, “Hamming distance,” https://en.wikipedia.org/wiki/Hamming_distance, 2019.
dc.identifier.urihttp://tdr.lib.ntu.edu.tw/jspui/handle/123456789/21591-
dc.description.abstract本論文提出一個在不藉助深度相機鏡頭、紅外線熱感應鏡頭等額外裝置,僅使用智慧型手機單顆可見光前鏡頭,辨識取得的影像來自真人使用者,或電子欺騙攻擊 (Spoofing Attack),包含紙本相片、電子相片、影片等。首先我們透過單樣本多框偵測器 (Single Shot Multi-Box Detector, SSD) 擷取畫面中的人臉,經過特定的資料擴增 (Data Augmentation) 後,再餵入捲積類神經網路 (Convolution Neural Network, CNN) ,訓練出具有分辨真人使用者與電子欺騙攻擊的模型。zh_TW
dc.description.abstractOur proposed method is capable of authenticating the input image is from real user or spoofing attack, including paper photograph, digital photograph, and video, using only the Red, Green, Blue (RGB) frontal camera of common smart phone, without the help of depth camera or infrared thermal sensor. We first capture live faces in each frame of input video streams by single shot multi-box detector then feed into our designed convolution neural network after certain data augmentation and finally obtain a well-trained spoof face classifier.en
dc.description.provenanceMade available in DSpace on 2021-06-08T03:39:00Z (GMT). No. of bitstreams: 1
ntu-108-R05945052-1.pdf: 3825847 bytes, checksum: d11275252f81bd4ceedca01eff97d694 (MD5)
Previous issue date: 2019
en
dc.description.tableofcontentsCONTENTS
誌謝 i
中文摘要 iii
ABSTRACT iv
CONTENTS v
LIST OF FIGURES ix
LIST OF TABLES xii
Chapter 1 Introduction 1
1.1 Overview 1
1.2 Acquisition of Training Images 3
1.3 Thesis Organization 4
Chapter 2 Related Work 5
2.1 Overview 5
2.2 Stereo Matching for Disparity and Depth 8
2.2.1 Local Methods 13
2.2.2 Global Methods 15
2.2.3 Semi-Global Methods 16
2.2.4 The Road Not Taken 16
Chapter 3 Background 17
3.1 Convolutional Neural Network 17
3.1.1 Architecture 18
3.1.2 Modeling one neuron 21
3.1.3 Strides 23
3.1.4 Pooling Layer 24
3.1.5 Convolution Layer 26
3.1.6 Fully Connected Layer 29
3.1.7 Activation Function 31
3.1.8 Softmax 33
3.2 Data Augmentation 34
3.2.1 Scaling 35
3.2.2 Translation 36
3.2.3 Rotation (at 90 Degrees): 36
3.2.4 Rotation (at Finer Angles): 36
3.2.5 Flipping: 38
3.2.6 Adding Salt and Pepper Noise: 39
3.2.7 Lighting Condition: 40
3.2.8 Perspective Transform: 40
Chapter 4 Methodology 42
4.1 Overview 42
4.2 Single Shot Multi-Box Detector (SSD) 43
4.2.1 Training 47
.4.2.1.1 Generate Priors 47
.4.2.1.2 MultiBox Detector 47
.4.2.1.3 Match Priors with Ground-Truth Boxes 48
.4.2.1.4 Scale Ground-Truth Boxes 50
.4.2.1.5 Hard Negative Mining 51
.4.2.1.6 The Loss Function 52
.4.2.1.7 Data Augmentation 52
4.2.2 Prediction 53
.4.2.2.1 NMS 54
4.2.3 The Drawbacks 56
4.3 Histogram Equalization 56
4.3.1 Histogram 56
4.3.2 Histogram Equalization 57
4.3.3 Adaptive Histogram Equalization 58
4.3.4 Contrastive Limited Adaptive Equalization 60
4.4 Data Augmentation 62
4.5 Neural Network 64
4.5.1 VGG-16 64
4.5.2 The architecture of neural network 66
Chapter 5 Experimental Result 72
5.1 Overview 72
5.2 Evaluation 73
5.3 Results and Observations 75
Chapter 6 Conclusion and Future Work 79
References 80
LIST OF FIGURES
Figure 2 1: RGB, depth map from depth camera and depth map from stereo matching. [6] 10
Figure 2 2: The notations and their relationship of stereo matching. [29] 12
Figure 3 1: Basic CNN structure. [21] 18
Figure 3 2: First layer of a convolutional neural network with pooling. Units of the same color have tied weights and units of different color represent different filter maps. [44] 20
Figure 3 3: Neuron unit. [39] 22
Figure 3 4: Stride of 2 pixels. [25] 23
Figure 3 5: An image max-pooled with 2x2 filters and stride 2. [25] 25
Figure 3 6: Convolution arithmetic. [13] 27
Figure 3 7: A 32x32 image padded with 2 pixels on each side. [13] 28
Figure 3 8: Some common filters. [25] 29
Figure 3 9: After pooling layer, flattened as FC layer. [25] 30
Figure 3 10: Complete CNN architecture. [25] 31
Figure 3 11: The pattern of a ReLU function. [22] 32
Figure 3 12: The pipeline of a softmax function. 33
Figure 3 13: Notice the background noise added in the image set. [31] 38
Figure 3 14: Salt and pepper noise added. [31] 39
Figure 3 15: Perspective transform. [31] 41
Figure 4 1: Pipeline of our proposed method. 42
Figure 4 2: VGG-based SSD architecture. [16] 46
Figure 4 3: SSD: Multiple Bounding Boxes for Localization (loc) and Confidence (conf). [43] 48
Figure 4 4: Diagram explaining IoU. [15] 49
Figure 4 5: Pseudo-code of Jaccard index. [16] 50
Figure 4 6: Example of hard negative mining. [15] 51
Figure 4 7: Some “Crazy” Detection Results on MS COCO Dataset. [15] 54
Figure 4 8: Pseudo code of NMS. 55
Figure 4 9: NMS example. [12] 55
Figure 4 10: The initial image and its histogram and cumulative histogram. [30] 57
Figure 4 11: The initial after histogram equalization. [30] 58
Figure 4 12: The comparison of initial image, histogram equalized image, and adaptive histogram equalized image. [30] 59
Figure 4 13: The initial image, histogram equalized image, and contrastive limited adaptive equalized image. [Sudhakar, 2019] 61
Figure 4 14: The initial image and contrastive limited adaptive equalized image. 62
Figure 4 15: Different VGG Layer Structures Using Single Scale (256) Evaluation. [42] 65
Figure 4 16: The architecture of our proposed neural network. 71
Figure 5 1: The training loss and accuracy on dataset. 73
Figure 5 2: Snapshots of testing results in true positive group. 75
Figure 5 4: Snapshots of testing results in true negative group. 76
Figure 5 5: Snapshots of testing results in miss detection group. 77
Figure 5 6: Snapshots of testing results in false alarm group. 78
LIST OF TABLES
Table 5 1: The environment to develop and conduct experiments is shown in this table. 72
Table 5 2: The testing result: percentage (number of frames / number of total frames). 74
dc.language.isoen
dc.title智慧型手機真人臉型驗證zh_TW
dc.titleAnti-Spoofing of Live Face Authentication on Smartphoneen
dc.typeThesis
dc.date.schoolyear107-2
dc.description.degree碩士
dc.contributor.oralexamcommittee江元麟,鄭宇哲,王獻章
dc.subject.keyword活臉辨識,電子欺騙攻擊,單樣本多框偵測器,資料擴增,捲積類神經網路,zh_TW
dc.subject.keywordLive Face Authentication,Spoofing Attack,Single Shot Multi-Box Detector,Data Augmentation,CNN,en
dc.relation.page86
dc.identifier.doi10.6342/NTU201901370
dc.rights.note未授權
dc.date.accepted2019-07-15
dc.contributor.author-college電機資訊學院zh_TW
dc.contributor.author-dept生醫電子與資訊學研究所zh_TW
顯示於系所單位:生醫電子與資訊學研究所

文件中的檔案:
檔案 大小格式 
ntu-108-1.pdf
  未授權公開取用
3.74 MBAdobe PDF
顯示文件簡單紀錄


系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。

社群連結
聯絡資訊
10617臺北市大安區羅斯福路四段1號
No.1 Sec.4, Roosevelt Rd., Taipei, Taiwan, R.O.C. 106
Tel: (02)33662353
Email: ntuetds@ntu.edu.tw
意見箱
相關連結
館藏目錄
國內圖書館整合查詢 MetaCat
臺大學術典藏 NTU Scholars
臺大圖書館數位典藏館
本站聲明
© NTU Library All Rights Reserved