請用此 Handle URI 來引用此文件:
http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/56362
完整後設資料紀錄
DC 欄位 | 值 | 語言 |
---|---|---|
dc.contributor.advisor | 洪一平(Yi-Ping Hung) | |
dc.contributor.author | Chih-Chuan Lai | en |
dc.contributor.author | 賴治權 | zh_TW |
dc.date.accessioned | 2021-06-16T05:25:08Z | - |
dc.date.available | 2014-08-21 | |
dc.date.copyright | 2014-08-21 | |
dc.date.issued | 2014 | |
dc.date.submitted | 2014-08-14 | |
dc.identifier.citation | [1] Applied Science Laboratories (ASL), Bedford, MA, USA. http://www.a-s-l.com.
[2] LC Technologies, Inc., Fairfax, VA, USA. http://www.eyegaze.com. [3] SensoMotoric Instruments GmbH (SMI), Teltow/Berlin, Germany. http://www.smivision.com. [4] Tobii Technology AB, Danderyd, Sweden. http://www.tobii.com. [5] Accurate two-dimensional eye tracker using first and fourth purkinje images. Journal of the Optical Society of America, 63(8):921–928, 1973. [6] J. Ahlberg. CANDIDE-3—An updated parameterized face. Dept. Elect. Eng., Linkoping Univ., Linkoping, Sweden, Tech. Rep., LiTH-ISY-R-2326, 2001. [7] S. Baluja and D. Pomerleau. Non-intrusive gaze tracking using artificial neural networks. Sch. Comput. Sci., Carnegie Mellon Univ., Pittsburgh, PA, USA, Tech. Rep., CMU-CS-94-102, 1994. [8] Florin B˘arbuceanu, Mihai Dugulean˘a, Stoianovici Vlad, and Adrian Nedelcu. Evaluation of the average selection speed ratio between an eye tracking and a head tracking interaction interface. Technological Innovation for Sustainability, pages 181–186, 2011. [9] D. Beymer and M. Flickner. Eye gaze tracking using an active stereo head. In International Conference on Computer vision and Pattern Recognition, pages 451–458, 2003. [10] L. Breiman. Random forests. Machine Learning, 45(1):5–32, 2001. [11] M. La Cascia, S. Sclaroff, and V. Athitsos. Fast, reliable head tracking under varying illumination: An approach based on registration of texture-mapped 3D model. IEEE Transactions on Pattern Analysis and Machine Intelligence, 22(4):322–336, 2000. [12] W. Y. Chang, C. S. Chen, and Y. D. Jian. Visual Tracking in High-Dimensional State Space by Appearance-Guided Particle Filtering. IEEE Transactions on Image Processing, 17(7):1154–1167, 2008. [13] T. F. Cootes and C. J.Taylor. Active Shape Models. In British Machine Vision Conference, pages 266–275, 1992. [14] T. Cornweet and H. Crane. Accurate two-dimensional eye tracker using first and fourth purkinje images. Journal of the Optical Society of America, 63(8):921–928, 1973. [15] G. Daunys and N. Ramanauskas. The accuracy of eye tracking using image processing. In Proceedings of the third Nordic conference on Human-computer interaction, pages 377–380, 2004. [16] D. F. DeMenthon and L. S. Davis. Model-based object pose in 25 lines of code. International Journal of Computer Vision, 15(1-2):123–141, 1995. [17] F. Dornaika and F. Davoine. On Appearance Based Face and Facial Action Tracking. IEEE Transactions on Circuits and Systems for Video Technology, 16(9):1107–1124, 2006. [18] M. Eizenman, L. H. Yu, L. Grupp, and M. Ellenbogen E. Eizenman, M. Gemar, and R. D. Levitan. A naturalistic visual scanning approach to assess selective attention in major depressive disorder. Psychiatry Research, 118(2):117–128, 2003. [19] G. Fanelli, T. Weise, J. Gall, and L. V. Gool. Real Time Head Pose Estimation from Consumer Depth Cameras. In Annual symposium of the German association for pattern recognition - DAGM 2011, pages 101–110, 2011. [20] M. A. Fischler and R. C. Bolles. Random sample consensus: A paradigm for model fiting with applications to image analysis and automated cartography. Communications of the ACM, 24(6):381–395, 1981. [21] A. W. Fitzgibbon, M. Pilu, and R. B. Fishe. Direct least square fitting of ellipses. IEEE Transactions on Pattern Analysis and Machine Intelligence, 21(5):476–480, 1999. [22] L. A. Frey, J. K. P. White, and T. E. Hutchinson. Eye-gaze word processing. IEEE Transactions on Systems, Man and Cybernetics, 20:944–950, 1990. [23] A. G. Gale. A note on the remote oculometer technique for recording eye movements. Vision Research, 22(1):201–202, 1982. [24] J. Gall and V. Lempitsky. Class-specific Hough forest for object detection. In Proceedings of IEEE Conference Computer Vision and Pattern Recognition, 2009. [25] J. H. Goldberg and A. M. Wichansky. Eye Tracking in Usability Evaluation: A Practitioner’s Guide. Elsevier Science, pages 493–516, 2003. [26] Y. Le Grand. Light, Color and Vision. 1957. New York: Wiley. [27] E .D. Guestrin and M. Eizenman. General theory of remote gaze estimation using the pupil center and corneal reflections. IEEE Transactions On Biomedical Engineering, 53(6):1124–1132, 2006. [28] S. R. Gulliver and G. Ghinea. Stars in their eyes: What eye-tracking reveals about multimedia perceptual quality. IEEE Transactions on Systems, Man and Cybernetics, Part A, 34(4):472–482, 2004. [29] Christian G‥utl, Maja Pivec, Christian Trummer, Victor Manuel Garc’ıa-Barrios, Felix M‥odritscher, Juergen Pripfl, and Martin Umgeher. Adele (adaptive e-learning with eye-tracking): Theoretical background, system architecture and application scenarios. Learning, 2012. [30] D.W. Hansen and Q. Ji. In the eye of the beholder: A survey of models for eyes and gaze. IEEE Transactions on Pattern Analysis and Machine Intelligence, 32(3):478–500, 2010. [31] R. Hartley and A. Zisserman. Multiple view geometry in computer vision, 2nd Edition. 2003. Cambridge University Press. [32] H. Hartridge and L. C. Thompson. Methods of investigating eye movements. British Journal of Ophthalmology, 32:581–591, 1948. [33] T. E. Hutchinson, J. K. P. White, W. M. Martin, K. C. Reichert, and L. A. Frey. Human-Computer Interaction Using Eye-Gaze Input. IEEE Transactions on Systems, Man and Cybernetics, 19:1527–1534, 1989. [34] A. D. Jepson. Robust online appearance models for visual tracking. IEEE Transactions on Pattern Analysis and Machine Intelligence, 25(10):1296–1311, 2003. [35] O. Jesorsky, K. J. Kirchberg, and R. Frischholz. Robust face detection using the Hausdorff distance. In Proceedings of Audio Video Biometric Person Authentication http://www.test.org/doe/, pages 90–95, 2001. [36] A. Kaufman, A. Bandopadhay, and B. Shaviv. An eye tracking computer user interface. In Proceedings of the Research Frontier in Virtual Reality Workshop, pages 78–84, 1993. [37] S. Kohlbecher, S. Bardins, K. Bartl, E. Schneider, T. Poitschke, and M. Ablassmeier. Calibration-free eye tracking by reconstruction of the pupil ellipse in 3D space. In Proceedings of Eye Tracking Research Applications Symposium, pages 135–138, 2008. [38] E. Konukoglu, B. Glocker, and D. Zikic. Neighbourhood Approximation Forests. In International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI), pages 75–82, 2012. [39] G. Loshe. Consumer eye movement patterns of Yellow Pages advertising. Journal of Advertising, 26:61–73, 1997. [40] F. Lu, Y. Sugano, T. Okabe, and Y. Sato. Inferring human gaze from appearance via adaptive linear regression. In In Proceedings of IEEE International Conference on Computer Vision, pages 153–160, 2011. [41] E. S. Rosen M. C. Corbett and D. P. S. O’Brart. Corneal Topography: Principles and Applications. 1999, p. 6. London, U.K.: BMJ Books. [42] S. D. Ma. Conics-based stereo, motion estimation, and pose determination. International Journal of Computer Vision, 10(1):7–25, 1993. [43] Y. Matsumoto and A. Zelinsky. An algorithm for real-time stereo vision implementation of head pose and gaze direction measurement. In International Conference Automatic Face Gesture Recognition, pages 499–504, 2000. [44] J. More. The levenberg-marquardt algorithm, implementation and theory. 1977. Springer-Verlag. [45] C. H. Morimoto and M. Mimica. Eye gaze tracking techniques for interactive applications. Computer Vision and Image Understanding, Special Issue on Eye Detection and Tracking, 98(1):4–24, 2005. [46] T. Ohno, N. Mukawa, and A. Yoshikawa. Freegaze: A gaze tracking system for everyday gaze interaction. In Proceedings Eye Tracking Research Applications Symposium, pages 125–132, 2002. [47] K. R. Park. A real-time gaze position estimation method based on a 3D eye model. IEEE Transactions on Systems, Man and Cybernetics, Part B, 37(1):199–212, 2007. [48] J C. Popieul, P. Simon, R. Leroux, and J C. Angu’e. Automatic processing of a car driver eye scanning movements on a simulated highway driving context. In in Proc. of the IEEE Intelligent Vehicles Symposium 2000, pages 13–18, 2000. [49] John P Rae, William Steptoe, and David J Roberts. Some implications of eye gaze behavior and perception for the design of immersive telecommunication systems. In Proceedings of the 2011 IEEE/ACM 15th International Symposium on Distributed Simulation and Real Time Applications, pages 108–114. IEEE Computer Society, 2011. [50] K. Rayner. Eye movements in reading and information processing: 20 years of research. Psychological Bulletin, 124:372–422, 1998. [51] D. A. Robinson. A method of measuring eye movements using a scleral search coil in a magnetic field. IEEE Transactions on Biomedical Engineering, 10:137–145, 1963. [52] R. Safaee-Rad, K. C. Smith, and B. Benhabib. Three-dimensional location estimation of circular features for machine vision. IEEE Transactions on Robotics and Automation, 8(5):624–640, 1992. [53] Y. Shan, Z. Liu, and Z. Zhang. Model-based bundle adjustment with application to face modeling. In Proceedings of IEEE International Conference on Computer Vision, pages 644–651, 2001. [54] R. Sharma, V. I. Pavlovi’c, and T. S. Huang. Toward multimodal human-computer interface. Proc. of the IEEE, 86(5):853–869, 1998. [55] S. W. Shih and J. Liu. A novel approach to 3-D gaze tracking using stereo cameras. IEEE Transactions on Systems, Man and Cybernetics, Part B, 34(1):234–245, 2004. [56] C. J. Stone. Consistent nonparametric regression. The Annals of Statistics, 5(4):595–620, 1977. [57] Y. Sugano, Y. Matsushita, Y. Sato, and H. Koike. An incremental learning method for unconstrained gaze estimation. In In Proceedings of the 10th European Conference on Computer Vision, pages 656–667, 2008. [58] K. H. Tan, D. J. Kriegman, and N. Ahuja. Appearance-based eye gaze estimation. In In Proceedings Sixth IEEE Workshop Applications of Computer Vision, pages 191–195, 2002. [59] G. Underwood. Cognitive Processes in Eye Guidance. 2005. Oxford Univ. Press. [60] A. Villanueva and R. Cabeza. Models for gaze tracking systems. Journal on Image and Video Processing, 2007(4):2:1–2:16, 2007. [61] A. Villanueva and R. Cabeza. Evaluation of corneal refraction in a model of a gaze tracking system. IEEE Transactions on Biomedical Engineering, 55(12):2812–2822, 2008. [62] P. Viola and M. Jones. Rapid object detection using a boosted cascade of simple features. In In Proceedings of IEEE Conference Computer Vision and Pattern Recognition, pages 511–518, 2002. [63] J. Wang and E. Sung. Study on eye gaze estimation. IEEE Transaction on Systems, Man, Cybernetics, Part B, 32(3):332–350, 2002. [64] P. A.Wetzel, G. Krueger-Anderson, C. Poprik, and P. Bascom. An eye tracking system for analysis of pilots’ scan paths. United States Air Force Armstrong Laboratory, Tech. Rep. AL/HR-TR-1996-0145, Apr. 1997. [65] O. M. C. Williams, A. Blake, and R. Cipolla. Sparse and semi-supervised visual mapping with the S3Gp. In In Proceedings of IEEE Conference Computer Vision and Pattern Recognition, pages 230–237, 2006. [66] J. Wright, A. Yang, A. Ganesh, S. Sastry, and Y. Ma. Robust face recognition via sparse representation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 31(2):210–227, 2008. [67] X. Xie and K. M. Lam. An efficient illumination normalization method for face recognition. Pattern Recognition Letters, 27(6):609–617, 2006. [68] D. H. Yoo and M. J. Chung. A novel non-intrusive eye gaze estimation using crossratio under large head motion. Computer Vision Image Understanding, 98(1):25–51, 2005. [69] L. R. Young and D. Sheena. Methods and designs-survey of eye movement recording methods. Behavior Research Methods Instrumentation, 7(5):397–429, 1975. [70] W. Zhang, Q. Wang, and X. Tang. Real Time Feature Based 3-D Deformable Face Tracking. In Proceedings of the European Conference on Computer Vision, pages 720–732, 2008. [71] Z. Zhang. A flexible new technique for camera calibration. IEEE Transactions on Pattern Analysis and Machine Intelligence, 22(11):1330–1334, 2000. [72] Z. Zhu and Q. Ji. Eye novel eye gaze tracking techniques under natural head movement. IEEE Transactions on Biomedical Engineering, 54(12):2246–2260, 2007. | |
dc.identifier.uri | http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/56362 | - |
dc.description.abstract | 視線追蹤技術在用來推論使用者所關注的焦點上一直扮演重要的角色。在論文中,我們發展兩個非接觸式三維視線追蹤的方法。一個是基於三維模型的視線追蹤方法;它使用立體視覺技術與高解析度的紅外線影像,適用於有高準確度要求的良好控制環境之下。另一個是以眼睛影像為基礎的視線追蹤方法;它使用市面上一般使用的攝影機且不需要額外的硬體設備,能夠運作在一個較寬鬆的環境但依然能提供一定範圍的準確度視線估測。
論文的第一部份提出一個基於三維模型的視線追蹤方法。角膜反射(或亮點)在三維模型的視線追蹤方法是一個很重要的特徵。然而當角膜反射在大視角範圍的使用下有亮點消失的問題,使得效能和使用的範圍因此受限。雖然瞳孔輪廓特徵提供相輔的資訊來幫助視線方位的計算,但現行的方法中並沒有在瞳孔輪廓特徵上提供一個正確角膜折射模型,因而導致不準確估測結果。我們提出一個以瞳孔輪廓特徵為基礎的視線追蹤方法,它能相容於角膜折射。此外我們也說明以亮點特徵與以瞳孔輪廓為特徵兩者方法的關連性,並將兩個方法整合在一個統一框架中。我們說明整合方法的必要性,結合兩個方法所提供的互補的資訊對於增強系統的穩定性與靈活性是有幫助的。 論文的第二部份發展一個以眼睛影像為基礎視線追蹤方法。直接使用眼睛影像作為基礎的視線追蹤方法的困難點在於攝影機所觀察到眼睛影像會隨著頭部方位而有所改變。為了克服這個問題,我們提出一個結合三維頭部方位追蹤與眼睛影像為基礎的視線追蹤方法。我們提出一個以三維人臉模型的頭部方位追蹤技術;它使用眼睛位置來做為引導輔助頭部追蹤,使得系統具有更穩固的特性;並發展一套自動初始化、追蹤失敗重新初始化的機制。我們基於霍夫隨機森林演算法在頭部方位空間與影像眼睛外觀的聯合空間建構出相鄰結構,然後在相鄰結構樣本中利用 l1 回歸法來找出最佳的視線估測。所建構的系統在使用者在頭部移動下能仍運作並提供一定準確度的三維視線估測。 | zh_TW |
dc.description.abstract | Gaze estimation techniques play important roles in inferring the focus of attention all the time. In this dissertation, we develop two remote 3-D gaze tracking methods. One is a
3-D model-based method that uses stereo vision with high resolution infrared imaging, which satisfies the high accuracy requirement for the use in controlled environments; the other is an appearance-based method that uses a commercial-off-the-shelf camera without requiring any dedicated hardware, which provides moderate estimation accuracy but can be used in less constrained environments. The first part of this thesis describes a 3-D model-based method for gaze tracking. Corneal reflection (or glint) is an important feature used by many 3-D model-based methods. However, when the operation range of a gaze tracking system is enlarged, glint feature-based (GFB) approaches will suffer from performance mainly due to the missing glint problem. Although the pupil contour feature may provide complementary information to help estimating the eye gaze, existing methods do not properly handle the cornea refraction problem which leads to inaccurate results. We describes a contour-feature based (CFB) 3-D gaze tracking method that is compatible to cornea refraction. We also show that both the GFB approach and the CFB approach can be formulated in a unified framework and, thus, they can be easily integrated. Furthermore, it is shown that the integration of the proposed CFB method and the GFB method is necessary because the two methods provide complementary information that helps to leverage the strength of both features and provides robustness and flexibility to the system. In the second part of this thesis, an appearance-based gaze tracking system is presented. The main difficulties for the appearance-based method are that eye appearances are varied with head motion. To overcome the difficulty, we propose a 3-D gaze tracking method combining head pose tracking and appearance-based gaze estimation. We present a robust model-based head pose tracking method guided by the eye location information and develop an automatic mechanism to deal with the problems of pose initialization, tracking loss recovery and re-initialization. We use a random forest approach to model the neighbor structure of the joint head pose and eye appearance space. l1-optimization is then used to seek for the best solution for regression from the selected neighboring samples. The constructed system is capable of analyzing the 3-D visual LoS of a person allowing the movement of his/her head and eye in a more natural manner but still provides moderate estimation accuracy. | en |
dc.description.provenance | Made available in DSpace on 2021-06-16T05:25:08Z (GMT). No. of bitstreams: 1 ntu-103-D94944004-1.pdf: 1438379 bytes, checksum: 003301dae2cc3783dc30bf2647ba9e88 (MD5) Previous issue date: 2014 | en |
dc.description.tableofcontents | Abstract vii
List of Figures xiii List of Tables xvii 1 Introduction 1 1.1 Background and Motivation . . . . . . . . . . . . . . . . . . . . . . . . 1 1.2 Human Eye Structure . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2 1.3 Cornea Reflection and Virtual Pupil . . . . . . . . . . . . . . . . . . . . 4 1.4 Overview of the Thesis . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 2 Related Work 7 2.1 Feature-based Method . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 2.1.1 2-D regression-based Approach . . . . . . . . . . . . . . . . . . 8 2.1.2 3-D model-based Approach . . . . . . . . . . . . . . . . . . . . 9 2.2 Appearance-based Method . . . . . . . . . . . . . . . . . . . . . . . . . 12 2.3 Contributions of Our Work . . . . . . . . . . . . . . . . . . . . . . . . . 14 3 3-D Model-based Method 17 3.1 Estimation of the EACC Plane . . . . . . . . . . . . . . . . . . . . . . . 18 3.1.1 Proposed CFB Approach for Estimating the EACC Plane . . . . . 18 3.1.2 Shih and Liu’s GFB Approach for Estimating the EACC Plane . . 24 3.1.3 Singular Cases and the Need for Integration . . . . . . . . . . . . 25 3.2 Estimation of the Line of Sight . . . . . . . . . . . . . . . . . . . . . . . 26 3.2.1 Estimation of the Optical Axis . . . . . . . . . . . . . . . . . . . 26 3.2.2 Inferring the Line of Sight Using the Optical Axis . . . . . . . . . 29 3.3 Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30 3.3.1 Necessity of Using the Contour Feature . . . . . . . . . . . . . . 30 3.3.2 Experiment Setup for Evaluating the Proposed Method . . . . . . 34 3.3.3 Experiments on Synthetic Data . . . . . . . . . . . . . . . . . . . 37 3.3.4 Experiments on Real Data . . . . . . . . . . . . . . . . . . . . . 42 3.4 Summary of the Proposed Model-based Method . . . . . . . . . . . . . . 49 4 Appearance-based Method 51 4.1 Problem statement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52 4.2 Eye Location Estimation . . . . . . . . . . . . . . . . . . . . . . . . . . 53 4.2.1 Training . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54 4.2.2 Eye Detection and Localization . . . . . . . . . . . . . . . . . . 56 4.3 Head Pose Tracking . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57 4.3.1 Deformable Face Model and Shape-free Facial Representation . . 57 4.3.2 Model Initialization . . . . . . . . . . . . . . . . . . . . . . . . . 59 4.3.3 Head Pose Tracking . . . . . . . . . . . . . . . . . . . . . . . . 61 4.4 Gaze Direction Estimation . . . . . . . . . . . . . . . . . . . . . . . . . 66 4.4.1 Neighbor Selection . . . . . . . . . . . . . . . . . . . . . . . . . 69 4.4.2 Gaze Direction Regression . . . . . . . . . . . . . . . . . . . . . 72 4.5 Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73 4.5.1 Evaluation of the Head Pose Estimation . . . . . . . . . . . . . . 74 4.5.2 Evaluation of the Gaze Estimation . . . . . . . . . . . . . . . . . 75 4.6 Summary of the Proposed Appearance-based Method . . . . . . . . . . . 78 5 Conclusions 81 | |
dc.language.iso | en | |
dc.title | 三維視線追蹤方法之研究 | zh_TW |
dc.title | Study on Methods of 3-D Gaze Tracking | en |
dc.type | Thesis | |
dc.date.schoolyear | 102-2 | |
dc.description.degree | 博士 | |
dc.contributor.oralexamcommittee | 傅立成(Li-Chen Fu),王傑智(Chieh-Chih (Bob),陳祝嵩(Chu-song Chen),范國清(Kuo-Chin Fan),賴尚宏(Shang-Hong Lai) | |
dc.subject.keyword | 視線追蹤,角膜反射,亮點,瞳孔輪廓,頭部方位追蹤, | zh_TW |
dc.subject.keyword | Gaze Tracking,Glint,Corneal Reflection,Pupil Contour,Head Pose Tracking, | en |
dc.relation.page | 90 | |
dc.rights.note | 有償授權 | |
dc.date.accepted | 2014-08-15 | |
dc.contributor.author-college | 電機資訊學院 | zh_TW |
dc.contributor.author-dept | 資訊網路與多媒體研究所 | zh_TW |
顯示於系所單位: | 資訊網路與多媒體研究所 |
文件中的檔案:
檔案 | 大小 | 格式 | |
---|---|---|---|
ntu-103-1.pdf 目前未授權公開取用 | 1.4 MB | Adobe PDF |
系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。