Skip navigation

DSpace

機構典藏 DSpace 系統致力於保存各式數位資料(如:文字、圖片、PDF)並使其易於取用。

點此認識 DSpace
DSpace logo
English
中文
  • 瀏覽論文
    • 校院系所
    • 出版年
    • 作者
    • 標題
    • 關鍵字
    • 指導教授
  • 搜尋 TDR
  • 授權 Q&A
    • 我的頁面
    • 接受 E-mail 通知
    • 編輯個人資料
  1. NTU Theses and Dissertations Repository
  2. 電機資訊學院
  3. 電子工程學研究所
請用此 Handle URI 來引用此文件: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/51283
完整後設資料紀錄
DC 欄位值語言
dc.contributor.advisor陳良基
dc.contributor.authorTse-En Pengen
dc.contributor.author彭則恩zh_TW
dc.date.accessioned2021-06-15T13:29:24Z-
dc.date.available2017-03-08
dc.date.copyright2016-03-08
dc.date.issued2016
dc.date.submitted2016-02-04
dc.identifier.citation[1] J. Aggarwal and L. Xia, Human activity recognition from 3d data: A
review,' Pattern Recognition Letters, vol. 48, pp. 70{80, 2014.
[2] M. Ryoo, Human activity prediction: Early recognition of ongoing
activities from streaming videos,' in Computer Vision (ICCV), 2011
IEEE International Conference on, pp. 1036{1043, IEEE, 2011.
[3] M. Raptis and L. Sigal, Poselet key-framing: A model for human
activity recognition,' in Computer Vision and Pattern Recognition
(CVPR), 2013 IEEE Conference on, pp. 2650{2657, IEEE, 2013.
[4] T. Lan, T.-C. Chen, and S. Savarese, A hierarchical representation for
future action prediction,' in Computer Vision–ECCV 2014, pp. 689{
704, Springer, 2014.
[5] B. J. Frey and D. Dueck, Clustering by passing messages between
data points,' science, vol. 315, no. 5814, pp. 972{976, 2007.
[6] L.-P. Morency, A. Quattoni, and T. Darrell, Latent-dynamic discrimi-
native models for continuous gesture recognition,' in Computer Vision
and Pattern Recognition, 2007. CVPR’07. IEEE Conference on, pp. 1{
8, IEEE, 2007.
[7] M. S. Ryoo and J. K. Aggarwal, UT-Interaction Dataset, ICPR
contest on Semantic Description of Human Activities (SDHA).'
http://cvrc.ece.utexas.edu/SDHA2010/Human Interaction.html,
2010.
[8] WillowGarage, Pr2 overview.' http://www.willowgarage.com/
pages/pr2/overview, 2013. [Online; accessed 20-May-2015].
[9] S. Schroeder, This is what google's self-driving car 'sees' as it makes a turn.' http://mashable.com/2013/05/03/
google-self-driving-car-sees/, 2013. [Online; accessed 20-May-2015].
[10] Wikipedia, Google driverless car.' http://en.wikipedia.org/wiki/
Google_driverless_car, 2013. [Online; accessed 20-May-2015].
[11] S. Edelstein, Bmw activeassist system lets self-driving cars get side-
ways and keep you on the road.' http://www.digitaltrends.com/
cars/bmw-activeassist-introduced-at-ces-2014/#!EkGS4, 2014.
[Online; accessed 20-May-2015].
[12] M. Munaro, F. Basso, and E. Menegatti, Tracking people within
groups with rgb-d data,' in Intelligent Robots and Systems (IROS),
2012 IEEE/RSJ International Conference on, pp. 2101{2107, 2012.
[13] J. Aggarwal and M. S. Ryoo, Human activity analysis: A review,'
ACM Computing Surveys (CSUR), vol. 43, no. 3, p. 16, 2011.
[14] I. Laptev, On space-time interest points,' International Journal of
Computer Vision, vol. 64, no. 2-3, pp. 107{123, 2005.
[15] A. Klaser, M. Marsza lek, and C. Schmid, A spatio-temporal descriptor
based on 3d-gradients,' in BMVC 2008-19th British Machine Vision
Conference, pp. 275{1, British Machine Vision Association, 2008.
[16] H. Wang, M. M. Ullah, A. Klaser, I. Laptev, and C. Schmid, Evalua-
tion of local spatio-temporal features for action recognition,' in BMVC
2009-British Machine Vision Conference, pp. 124{1, BMVA Press,
2009.
[17] I. Laptev, M. Marsza lek, C. Schmid, and B. Rozenfeld, Learning re-
alistic human actions from movies,' in Computer Vision and Pattern
Recognition, 2008. CVPR 2008. IEEE Conference on, pp. 1{8, IEEE,
2008.
[18] L. Wang, Y. Qiao, and X. Tang, Motionlets: Mid-level 3d parts for
human motion recognition,' in Proceedings of the IEEE Conference on
Computer Vision and Pattern Recognition, pp. 2674{2681, 2013.
[19] E. H. Adelson and J. R. Bergen, Spatiotemporal energy models for
the perception of motion,' JOSA A, vol. 2, no. 2, pp. 284{299, 1985.
[20] S. Sadanand and J. J. Corso, Action bank: A high-level representa-
tion of activity in video,' in Computer Vision and Pattern Recognition
(CVPR), 2012 IEEE Conference on, pp. 1234{1241, IEEE, 2012.
[21] K. G. Derpanis, M. Sizintsev, K. Cannons, and R. P. Wildes, Effi-
cient action spotting based on a spacetime oriented structure represen-
tation,' in Computer Vision and Pattern Recognition (CVPR), 2010
IEEE Conference on, pp. 1990{1997, IEEE, 2010.
[22] W. T. Freeman and E. H. Adelson, The design and use of steerable filters,' IEEE Transactions on Pattern Analysis & Machine Intelligence,
no. 9, pp. 891{906, 1991.
[23] P. Doll ar, V. Rabaud, G. Cottrell, and S. Belongie, Behavior recog-
nition via sparse spatio-temporal features,' in Visual Surveillance and
Performance Evaluation of Tracking and Surveillance, 2005. 2nd Joint
IEEE International Workshop on, pp. 65{72, IEEE, 2005.
[24] S. B. Wang and D. Demirdjian, Inferring body pose using speech con-
tent,' in Proceedings of the 7th international conference on Multimodal
interfaces, pp. 53{60, ACM, 2005.
[25] S. Fujie, Y. Ejiri, K. Nakajima, Y. Matsusaka, and T. Kobayashi, A
conversation robot using head gesture recognition as para-linguistic
information,' in Robot and Human Interactive Communication, 2004.
ROMAN 2004. 13th IEEE International Workshop on, pp. 159{164,
IEEE, 2004.
[26] J. Lafferty, A. McCallum, and F. C. Pereira, Conditional random filelds: Probabilistic models for segmenting and labeling sequence data,'
2001.
[27] A. Quattoni, M. Collins, and T. Darrell, Conditional random fields
for object recognition,' in Advances in neural information processing
systems, pp. 1097{1104, 2004.
[28] C. Cortes and V. Vapnik, Support-vector networks,' Machine learn-
ing, vol. 20, no. 3, pp. 273{297, 1995.
dc.identifier.urihttp://tdr.lib.ntu.edu.tw/jspui/handle/123456789/51283-
dc.description.abstractThe ultimate goal of computer vision is to help computing devices understand the real world, process visual information efficiently, and even have semantic understandings like humans do. Nowadays, computer vision algorithms progressed rapidly, and developed plenty innovative applications. For example, intelligent environmental surveillances of the future are capable of monitoring real environments, including objects and people.
Rather than still images, videos including spatial-temporal information imply richer knowledge. Therefore, human action recognition becomes a basic application that can be implemented in the vision of robots. The fact that different variations in videos increases the difficulty of analysis, leading many researchers to develop better algorithms aiming at raising the recognition accuracy on datasets. However, the computation complexity of feature extraction and template matching in videos is still too complicated to be real-time in past researches.
In the thesis, we first introduce several applications of computer vision. Then, we introduce the challenge and background knowledge of the action prediction system. Furthermore, we review the related algorithms and proposed a novel learning scheme for action prediction system. Last, we adapt our prediction system for real-world streaming scenario and explore the hardware-oriented optimization for such system.
en
dc.description.provenanceMade available in DSpace on 2021-06-15T13:29:24Z (GMT). No. of bitstreams: 1
ntu-105-R02943034-1.pdf: 8579188 bytes, checksum: 78180a652c8dbe09a7869179302019ad (MD5)
Previous issue date: 2016
en
dc.description.tableofcontentsAbstract
xiii
1 Introduction
1
1.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Applications of Computer Vision . . . . . . . . . . . . . . . 1
1.3 Motivation of Action Prediction . . . . . . . . . . . . . . . . 3
1.4 Design Considerations and Main Contributions . . . . . . . . 4
1.5 Thesis Organization . . . . . . . . . . . . . . . . . . . . . . . 5
2 Analysis of Action Prediction System
7
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
2.2 Challenges of Action Prediction System . . . . . . . . . . . . 8
2.3 Basics of Action Prediction System . . . . . . . . . . . . . . 9
2.3.1 Feature Feature Representation . . . . . . . . . . . . 9
2.3.2 Learning Models . . . . . . . . . . . . . . . . . . . . 10
2.4 Related Work . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
3 Proposed Part-based Action Prediction System 15
3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
3.2 System Overview . . . . . . . . . . . . . . . . . . . . . . . . 17
3.3 Feature Extraction . . . . . . . . . . . . . . . . . . . . . . . 19
3.3.1 SOE . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
3.3.2 Dense HOE and HOG . . . . . . . . . . . . . . . . . 21
3.4 Automatic Part Learning Scheme . . . . . . . . . . . . . . . 22
3.4.1 Saliency Region Extraction . . . . . . . . . . . . . . . 22
3.4.2 Exemplar Based Clustering . . . . . . . . . . . . . . 27
3.4.3 Parts Selection . . . . . . . . . . . . . . . . . . . . . 34
3.5 Prediction System . . . . . . . . . . . . . . . . . . . . . . . . 36
3.6 Experiment Results . . . . . . . . . . . . . . . . . . . . . . . 39
3.7 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
4 Proposed On-the-
y Prediction System and Architecture
Design
45
4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
4.2 On-the-
y Prediction system . . . . . . . . . . . . . . . . . . 45
4.3 Experiment Results . . . . . . . . . . . . . . . . . . . . . . . 47
4.4 Algorithm Optimization and Architecture Design . . . . . . 49
4.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
5 Conclusion
55
Bibliography
57
dc.language.isoen
dc.subject即時動作預測系統zh_TW
dc.subject自動部位學習方法zh_TW
dc.subjectOn-the-fly action prediction frameworken
dc.subjectautomatic part learning scheme.en
dc.title即時人類動作預測之演算法與架構設計zh_TW
dc.titleAlgorithm and Architecture Design for On-the-fly Human Action Predictionen
dc.typeThesis
dc.date.schoolyear104-1
dc.description.degree碩士
dc.contributor.oralexamcommittee賴永康,陳美娟,黃朝宗
dc.subject.keyword即時動作預測系統,自動部位學習方法,zh_TW
dc.subject.keywordOn-the-fly action prediction framework,automatic part learning scheme.,en
dc.relation.page60
dc.rights.note有償授權
dc.date.accepted2016-02-04
dc.contributor.author-college電機資訊學院zh_TW
dc.contributor.author-dept電子工程學研究所zh_TW
顯示於系所單位:電子工程學研究所

文件中的檔案:
檔案 大小格式 
ntu-105-1.pdf
  未授權公開取用
8.38 MBAdobe PDF
顯示文件簡單紀錄


系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。

社群連結
聯絡資訊
10617臺北市大安區羅斯福路四段1號
No.1 Sec.4, Roosevelt Rd., Taipei, Taiwan, R.O.C. 106
Tel: (02)33662353
Email: ntuetds@ntu.edu.tw
意見箱
相關連結
館藏目錄
國內圖書館整合查詢 MetaCat
臺大學術典藏 NTU Scholars
臺大圖書館數位典藏館
本站聲明
© NTU Library All Rights Reserved