請用此 Handle URI 來引用此文件:
http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/51283完整後設資料紀錄
| DC 欄位 | 值 | 語言 |
|---|---|---|
| dc.contributor.advisor | 陳良基 | |
| dc.contributor.author | Tse-En Peng | en |
| dc.contributor.author | 彭則恩 | zh_TW |
| dc.date.accessioned | 2021-06-15T13:29:24Z | - |
| dc.date.available | 2017-03-08 | |
| dc.date.copyright | 2016-03-08 | |
| dc.date.issued | 2016 | |
| dc.date.submitted | 2016-02-04 | |
| dc.identifier.citation | [1] J. Aggarwal and L. Xia, Human activity recognition from 3d data: A
review,' Pattern Recognition Letters, vol. 48, pp. 70{80, 2014. [2] M. Ryoo, Human activity prediction: Early recognition of ongoing activities from streaming videos,' in Computer Vision (ICCV), 2011 IEEE International Conference on, pp. 1036{1043, IEEE, 2011. [3] M. Raptis and L. Sigal, Poselet key-framing: A model for human activity recognition,' in Computer Vision and Pattern Recognition (CVPR), 2013 IEEE Conference on, pp. 2650{2657, IEEE, 2013. [4] T. Lan, T.-C. Chen, and S. Savarese, A hierarchical representation for future action prediction,' in Computer Vision–ECCV 2014, pp. 689{ 704, Springer, 2014. [5] B. J. Frey and D. Dueck, Clustering by passing messages between data points,' science, vol. 315, no. 5814, pp. 972{976, 2007. [6] L.-P. Morency, A. Quattoni, and T. Darrell, Latent-dynamic discrimi- native models for continuous gesture recognition,' in Computer Vision and Pattern Recognition, 2007. CVPR’07. IEEE Conference on, pp. 1{ 8, IEEE, 2007. [7] M. S. Ryoo and J. K. Aggarwal, UT-Interaction Dataset, ICPR contest on Semantic Description of Human Activities (SDHA).' http://cvrc.ece.utexas.edu/SDHA2010/Human Interaction.html, 2010. [8] WillowGarage, Pr2 overview.' http://www.willowgarage.com/ pages/pr2/overview, 2013. [Online; accessed 20-May-2015]. [9] S. Schroeder, This is what google's self-driving car 'sees' as it makes a turn.' http://mashable.com/2013/05/03/ google-self-driving-car-sees/, 2013. [Online; accessed 20-May-2015]. [10] Wikipedia, Google driverless car.' http://en.wikipedia.org/wiki/ Google_driverless_car, 2013. [Online; accessed 20-May-2015]. [11] S. Edelstein, Bmw activeassist system lets self-driving cars get side- ways and keep you on the road.' http://www.digitaltrends.com/ cars/bmw-activeassist-introduced-at-ces-2014/#!EkGS4, 2014. [Online; accessed 20-May-2015]. [12] M. Munaro, F. Basso, and E. Menegatti, Tracking people within groups with rgb-d data,' in Intelligent Robots and Systems (IROS), 2012 IEEE/RSJ International Conference on, pp. 2101{2107, 2012. [13] J. Aggarwal and M. S. Ryoo, Human activity analysis: A review,' ACM Computing Surveys (CSUR), vol. 43, no. 3, p. 16, 2011. [14] I. Laptev, On space-time interest points,' International Journal of Computer Vision, vol. 64, no. 2-3, pp. 107{123, 2005. [15] A. Klaser, M. Marsza lek, and C. Schmid, A spatio-temporal descriptor based on 3d-gradients,' in BMVC 2008-19th British Machine Vision Conference, pp. 275{1, British Machine Vision Association, 2008. [16] H. Wang, M. M. Ullah, A. Klaser, I. Laptev, and C. Schmid, Evalua- tion of local spatio-temporal features for action recognition,' in BMVC 2009-British Machine Vision Conference, pp. 124{1, BMVA Press, 2009. [17] I. Laptev, M. Marsza lek, C. Schmid, and B. Rozenfeld, Learning re- alistic human actions from movies,' in Computer Vision and Pattern Recognition, 2008. CVPR 2008. IEEE Conference on, pp. 1{8, IEEE, 2008. [18] L. Wang, Y. Qiao, and X. Tang, Motionlets: Mid-level 3d parts for human motion recognition,' in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2674{2681, 2013. [19] E. H. Adelson and J. R. Bergen, Spatiotemporal energy models for the perception of motion,' JOSA A, vol. 2, no. 2, pp. 284{299, 1985. [20] S. Sadanand and J. J. Corso, Action bank: A high-level representa- tion of activity in video,' in Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on, pp. 1234{1241, IEEE, 2012. [21] K. G. Derpanis, M. Sizintsev, K. Cannons, and R. P. Wildes, Effi- cient action spotting based on a spacetime oriented structure represen- tation,' in Computer Vision and Pattern Recognition (CVPR), 2010 IEEE Conference on, pp. 1990{1997, IEEE, 2010. [22] W. T. Freeman and E. H. Adelson, The design and use of steerable filters,' IEEE Transactions on Pattern Analysis & Machine Intelligence, no. 9, pp. 891{906, 1991. [23] P. Doll ar, V. Rabaud, G. Cottrell, and S. Belongie, Behavior recog- nition via sparse spatio-temporal features,' in Visual Surveillance and Performance Evaluation of Tracking and Surveillance, 2005. 2nd Joint IEEE International Workshop on, pp. 65{72, IEEE, 2005. [24] S. B. Wang and D. Demirdjian, Inferring body pose using speech con- tent,' in Proceedings of the 7th international conference on Multimodal interfaces, pp. 53{60, ACM, 2005. [25] S. Fujie, Y. Ejiri, K. Nakajima, Y. Matsusaka, and T. Kobayashi, A conversation robot using head gesture recognition as para-linguistic information,' in Robot and Human Interactive Communication, 2004. ROMAN 2004. 13th IEEE International Workshop on, pp. 159{164, IEEE, 2004. [26] J. Lafferty, A. McCallum, and F. C. Pereira, Conditional random filelds: Probabilistic models for segmenting and labeling sequence data,' 2001. [27] A. Quattoni, M. Collins, and T. Darrell, Conditional random fields for object recognition,' in Advances in neural information processing systems, pp. 1097{1104, 2004. [28] C. Cortes and V. Vapnik, Support-vector networks,' Machine learn- ing, vol. 20, no. 3, pp. 273{297, 1995. | |
| dc.identifier.uri | http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/51283 | - |
| dc.description.abstract | The ultimate goal of computer vision is to help computing devices understand the real world, process visual information efficiently, and even have semantic understandings like humans do. Nowadays, computer vision algorithms progressed rapidly, and developed plenty innovative applications. For example, intelligent environmental surveillances of the future are capable of monitoring real environments, including objects and people.
Rather than still images, videos including spatial-temporal information imply richer knowledge. Therefore, human action recognition becomes a basic application that can be implemented in the vision of robots. The fact that different variations in videos increases the difficulty of analysis, leading many researchers to develop better algorithms aiming at raising the recognition accuracy on datasets. However, the computation complexity of feature extraction and template matching in videos is still too complicated to be real-time in past researches. In the thesis, we first introduce several applications of computer vision. Then, we introduce the challenge and background knowledge of the action prediction system. Furthermore, we review the related algorithms and proposed a novel learning scheme for action prediction system. Last, we adapt our prediction system for real-world streaming scenario and explore the hardware-oriented optimization for such system. | en |
| dc.description.provenance | Made available in DSpace on 2021-06-15T13:29:24Z (GMT). No. of bitstreams: 1 ntu-105-R02943034-1.pdf: 8579188 bytes, checksum: 78180a652c8dbe09a7869179302019ad (MD5) Previous issue date: 2016 | en |
| dc.description.tableofcontents | Abstract
xiii 1 Introduction 1 1.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1.2 Applications of Computer Vision . . . . . . . . . . . . . . . 1 1.3 Motivation of Action Prediction . . . . . . . . . . . . . . . . 3 1.4 Design Considerations and Main Contributions . . . . . . . . 4 1.5 Thesis Organization . . . . . . . . . . . . . . . . . . . . . . . 5 2 Analysis of Action Prediction System 7 2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 2.2 Challenges of Action Prediction System . . . . . . . . . . . . 8 2.3 Basics of Action Prediction System . . . . . . . . . . . . . . 9 2.3.1 Feature Feature Representation . . . . . . . . . . . . 9 2.3.2 Learning Models . . . . . . . . . . . . . . . . . . . . 10 2.4 Related Work . . . . . . . . . . . . . . . . . . . . . . . . . . 10 2.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 3 Proposed Part-based Action Prediction System 15 3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 3.2 System Overview . . . . . . . . . . . . . . . . . . . . . . . . 17 3.3 Feature Extraction . . . . . . . . . . . . . . . . . . . . . . . 19 3.3.1 SOE . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 3.3.2 Dense HOE and HOG . . . . . . . . . . . . . . . . . 21 3.4 Automatic Part Learning Scheme . . . . . . . . . . . . . . . 22 3.4.1 Saliency Region Extraction . . . . . . . . . . . . . . . 22 3.4.2 Exemplar Based Clustering . . . . . . . . . . . . . . 27 3.4.3 Parts Selection . . . . . . . . . . . . . . . . . . . . . 34 3.5 Prediction System . . . . . . . . . . . . . . . . . . . . . . . . 36 3.6 Experiment Results . . . . . . . . . . . . . . . . . . . . . . . 39 3.7 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43 4 Proposed On-the- y Prediction System and Architecture Design 45 4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . 45 4.2 On-the- y Prediction system . . . . . . . . . . . . . . . . . . 45 4.3 Experiment Results . . . . . . . . . . . . . . . . . . . . . . . 47 4.4 Algorithm Optimization and Architecture Design . . . . . . 49 4.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52 5 Conclusion 55 Bibliography 57 | |
| dc.language.iso | en | |
| dc.subject | 即時動作預測系統 | zh_TW |
| dc.subject | 自動部位學習方法 | zh_TW |
| dc.subject | On-the-fly action prediction framework | en |
| dc.subject | automatic part learning scheme. | en |
| dc.title | 即時人類動作預測之演算法與架構設計 | zh_TW |
| dc.title | Algorithm and Architecture Design for On-the-fly Human Action Prediction | en |
| dc.type | Thesis | |
| dc.date.schoolyear | 104-1 | |
| dc.description.degree | 碩士 | |
| dc.contributor.oralexamcommittee | 賴永康,陳美娟,黃朝宗 | |
| dc.subject.keyword | 即時動作預測系統,自動部位學習方法, | zh_TW |
| dc.subject.keyword | On-the-fly action prediction framework,automatic part learning scheme., | en |
| dc.relation.page | 60 | |
| dc.rights.note | 有償授權 | |
| dc.date.accepted | 2016-02-04 | |
| dc.contributor.author-college | 電機資訊學院 | zh_TW |
| dc.contributor.author-dept | 電子工程學研究所 | zh_TW |
| 顯示於系所單位: | 電子工程學研究所 | |
文件中的檔案:
| 檔案 | 大小 | 格式 | |
|---|---|---|---|
| ntu-105-1.pdf 未授權公開取用 | 8.38 MB | Adobe PDF |
系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。
