Skip navigation

DSpace

機構典藏 DSpace 系統致力於保存各式數位資料(如:文字、圖片、PDF)並使其易於取用。

點此認識 DSpace
DSpace logo
English
中文
  • 瀏覽論文
    • 校院系所
    • 出版年
    • 作者
    • 標題
    • 關鍵字
    • 指導教授
  • 搜尋 TDR
  • 授權 Q&A
    • 我的頁面
    • 接受 E-mail 通知
    • 編輯個人資料
  1. NTU Theses and Dissertations Repository
  2. 電機資訊學院
  3. 電機工程學系
請用此 Handle URI 來引用此文件: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/87342
完整後設資料紀錄
DC 欄位值語言
dc.contributor.advisor傅立成zh_TW
dc.contributor.advisorLi-Chen Fuen
dc.contributor.author林子涵zh_TW
dc.contributor.authorTzu-Han Linen
dc.date.accessioned2023-05-18T17:10:33Z-
dc.date.available2023-11-10-
dc.date.copyright2023-06-20-
dc.date.issued2023-
dc.date.submitted2023-02-05-
dc.identifier.citation[1] Ifr., “Service Robots Record: Sales Worldwide up 32%.” IFR International Federation of Robotics, https://ifr.org/ifr-press-releases/news/service-robots-record-sales-worldwide-up-32.
[2] Mueller, M., Kuhlenkoetter, B., & Nassmacher, R., “Robots in Food Industry Challenges and Chances: Semantic Scholar.” Undefined, 1 Jan. 1970, https://www.semanticscholar.org/paper/Robots-in-Food-Industry-Challenges-and-Chances-Mueller-Kuhlenkoetter/2197ce8501badc66f645267bc84a9d9fb4f34af0.
[3] Kar, Suman, “Robotics in Healthcare.” 2019 2nd International Conference on Power Energy, Environment and Intelligent Control (PEEIC), 2019, https://doi.org/10.1109/peeic47157.2019.8976668.
[4] Samala, N., Katkam, B. S., Bellamkonda, R. S., & Rodriguez, R. V., “Impact of AI and Robotics in the Tourism Sector: A Critical Insight.” Journal of Tourism Futures, vol. 8, no. 1, 2020, pp. 73–87., https://doi.org/10.1108/jtf-07-2019-0065.
[5] “Robotic Vacuum Cleaner.” Wikipedia, Wikimedia Foundation, 15 July 2022, https://en.wikipedia.org/wiki/Robotic_vacuum_cleaner.
[6] G. Grisetti Dept. of Comput. Sci., Grisetti, G., Sci., D. of C., Stachniss, C., Profile, V., Burgard, W., & Metrics, O. M. V. A., “Improved Techniques for Grid Mapping with Rao-Blackwellized Particle Filters.” IEEE Transactions on Robotics, vol. 23, no. 1, 2007, pp. 34–46.
[7] Rösmann, C., Hoffmann, F., & Bertram, T., “Integrated Online Trajectory Planning and Optimization in Distinctive Topologies.” Robotics and Autonomous Systems, vol. 88, 2017, pp. 142–153.
[8] Granada, R., Monteiro, J., Gavenski, N., & Meneguzzi, F., “Object-Based Goal Recognition Using Real-World Data.” Advances in Soft Computing, 2020, pp. 325–337.
[9] Mekruksavanich, S., Promsakon, C., & Jitpattanakul, A., “Location-Based Daily Human Activity Recognition Using Hybrid Deep Learning Network.” 2021 18th International Joint Conference on Computer Science and Software Engineering (JCSSE), 2021.
[10] Snehitha, B., Sreeya, R. S., & Manikandan, V. M., “Human Activity Detection from Still Images Using Deep Learning Techniques.” 2021 International Conference on Control, Automation, Power and Signal Processing (CAPS), 2021.
[11] Babiker, M., Khalifa, O. O., Htike, K. K., Hassan, A., & Zaharadeen, M., “Automated Daily Human Activity Recognition for Video Surveillance Using Neural Network.” 2017 IEEE 4th International Conference on Smart Instrumentation, Measurement and Application (ICSIMA), 2017.
[12] Archana, N., and K. Hareesh., “Real-Time Human Activity Recognition Using ResNet and 3d Convolutional Neural Networks.” 2021 2nd International Conference on Advances in Computing, Communication, Embedded and Secure Systems (ACCESS), 2021.
[13] Song, S., Lan, C., Xing, J., Zeng, W., & Liu, J., “An End-to-End Spatio-Temporal Attention Model for Human Action Recognition from Skeleton Data.” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31, no. 1, 2017.
[14] Zhang, P., Lan, C., Xing, J., Zeng, W., Xue, J., & Zheng, N., “View Adaptive Recurrent Neural Networks for High Performance Human Action Recognition from Skeleton Data.” 2017 IEEE International Conference on Computer Vision (ICCV), 2017.
[15] Shi, L., Zhang, Y., Cheng, J., & Lu, H., “Non-Local Graph Convolutional Networks for Skeleton-Based Action Recognition.” ArXiv.org, 4 July 2018, https://arxiv.org/abs/1805.07694v2.
[16] Ren, B., Liu, M., Ding, R., & Liu, H., “A Survey on 3D Skeleton-Based Action Recognition Using Learning Method.” ArXiv.org, 14 Feb. 2020, https://arxiv.org/abs/2002.05907.
[17] Ke, Q., Bennamoun, M., An, S., Sohel, F., & Boussaid, F., “A New Representation of Skeleton Sequences for 3D Action Recognition.” 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
[18] Bagate, Asmita, & Medha Shah., “Human Activity Recognition Using RGB-D Sensors.” 2019 International Conference on Intelligent Computing and Control Systems (ICCS), 2019.
[19] Song, Y.-F., Zhang, Z., Shan, C., & Wang, L., “Richly Activated Graph Convolutional Network for Robust Skeleton-Based Action Recognition.” IEEE Transactions on Circuits and Systems for Video Technology, vol. 31, no. 5, 2021, pp. 1915–1925.
[20] Heidari, Negar, & Alexandros Iosifidis., “Temporal Attention-Augmented Graph Convolutional Network for Efficient Skeleton-Based Human Action Recognition.” 2020 25th International Conference on Pattern Recognition (ICPR), 2021.
[21] Li, S., Yi, J., Farha, Y. A., & Gall, J., “Pose Refinement Graph Convolutional Network for Skeleton-Based Action Recognition.” IEEE Robotics and Automation Letters, vol. 6, no. 2, 2021, pp. 1028–1035.
[22] Wang, H., Yu, B., Li, J., Zhang, L., & Chen, D., “Multi-Stream Interaction Networks for Human Action Recognition.” IEEE Transactions on Circuits and Systems for Video Technology, vol. 32, no. 5, 2022, pp. 3050–3060.
[23] Wu, Z., Pan, S., Chen, F., Long, G., Zhang, C., & Yu, P. S., “A Comprehensive Survey on Graph Neural Networks.” IEEE Transactions on Neural Networks and Learning Systems, vol. 32, no. 1, 2021, pp. 4–24.
[24] Jang, J., Kim, D., Park, C., Jang, M., Lee, J., & Kim, J., “ETRI-Activity3D: A Large-Scale RGB-D Dataset for Robots to Recognize Daily Activities of the Elderly.” 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2020.
[25] Van-Horenbeke, Franz A., & Angelika Peer., “Activity, Plan, and Goal Recognition: A Review.” Frontiers in Robotics and AI, vol. 8, 2021.
[26] Avrahami-Zilberbrand, D., Kaminka, G.A., & Zarosim, H., “Fast and Complete Symbolic Plan Recognition: Allowing for Duration, Interleaved Execution, and Lossy Observations.” International Joint Conference on Artificial intelligence (IJCAI), 2005.
[27] Geib, Christopher W., & Robert P. Goldman., “A Probabilistic Plan Recognition Algorithm Based on Plan Tree Grammars.” Artificial Intelligence, vol. 173, no. 11, 2009, pp. 1101–1132.
[28] Marchetta, M. G., & R. Q. Forradellas., “A New Model for Automatic Generation of Plan Libraries for Plan Recognition”. Brazilian Journal of Operations & Production Management, vol. 3, no. 2, Feb. 2010, pp. 5-19, https://bjopm.org.br/bjopm/article/view/BJV3N2_2006_P1.
[29] Geib, C.W., Maraist, J., & Goldman, R.P., “A New Probabilistic Plan Recognition Algorithm Based on String Rewriting.” ICAPS, 2008.
[30] Granada, R.L., Pereira, R., Monteiro, J., Barros, R.C., Ruiz, D.D., & Meneguzzi, F., “Hybrid Activity and Plan Recognition for Video Streams.” AAAI Workshops, 2017.
[31] Shimada, A., Kondo, K., Deguchi, D., Morin, G., & Stern, H., “Kitchen Scene Context Based Gesture Recognition: A Contest in ICPR2012.” Advances in Depth Image Analysis and Applications, 2013, pp. 168–185.
[32] Meng, L., Miao, C., & Leung, C., “Towards Online and Personalized Daily Activity Recognition, Habit Modeling, and Anomaly Detection for the Solitary Elderly through Unobtrusive Sensing.” Multimedia Tools and Applications, vol. 76, no. 8, 2016, pp. 10779–10799.
[33] Zhuo, Hankz., “Human-Aware Plan Recognition.” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31, no. 1, 2017.
[34] “Adjacency Matrix.” Wikipedia, Wikimedia Foundation, 7 June 2022, https://en.wikipedia.org/wiki/Adjacency_matrix.
[35] “Degree Matrix.” Wikipedia, Wikimedia Foundation, 26 Oct. 2021, https://en.wikipedia.org/wiki/Degree_matrix.
[36] “Laplacian Matrix.” Wikipedia, Wikimedia Foundation, 11 June 2022, https://en.wikipedia.org/wiki/Laplacian_matrix.
[37] J. Bruna, W. Zaremba, A. Szlam, Y.L. Cun, “Spectral Networks and Deep Locally Connected Networks on Graphs,” International Conference on Learning Representations (ICLR), 2014.
[38] Kipf, T. N., & Welling, M., “Semi-Supervised Classification With Graph Convolutional Networks,” International Conference on Learning Representations (ICLR), 2017.
[39] Yan, S., Xiong, Y., & Lin, D., “Spatial Temporal Graph Convolutional Networks for Skeleton-Based Action Recognition,” Association for the Advancement of AI (AAAI), 2018.
[39] Yan, S., Xiong, Y., & Lin, D., “Spatial Temporal Graph Convolutional Networks for Skeleton-Based Action Recognition,” Association for the Advancement of AI (AAAI), 2018.
[40] Shi, L., Zhang, Y., Cheng, J., & Lu, H., “Two-Stream Adaptive Graph Convolutional Networks for Skeleton-Based Action Recognition,” Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
[41] Girshick, R., Donahue, J., Darrell, T., & Malik, J., “Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation.” 2014 IEEE Conference on Computer Vision and Pattern Recognition, 2014.
[42] Girshick, Ross., “Fast R-CNN.” 2015 IEEE International Conference on Computer Vision (ICCV), 2015.
[43] Ren, S., He, K., Girshick, R., & Sun, J., “Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 6, 2017, pp. 1137–1149.
[44] C. Liu, Y. Tao, J. Liang, K. Li & Y. Chen., “Object Detection Based on Yolo Network.” 2018 IEEE 4th Information Technology and Mechatronics Engineering Conference (ITOEC), 2018.
[45] Shuai, Qianjun, & Xingwen Wu., “Object Detection System Based on SSD Algorithm.” 2020 International Conference on Culture-Oriented Science & Technology (ICCST), 2020.
[46] Del Prete, R., Graziano, M. D., & Renga, A., “RetinaNet: A Deep Learning Architecture to Achieve a Robust Wake Detector in SAR Images.” 2021 IEEE 6th International Forum on Research and Technology for Society and Industry (RTSI), 2021.
[47] Redmon, Joseph, & Ali Farhadi., “Yolo9000: Better, Faster, Stronger.” 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
[48] Li, S., Tao, F., Shi, T., & Kuang, J., “Improvement of yolov3 Network Based on Roi.” 2019 IEEE 4th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), 2019.
[49] J. Redmon, S. Divvala, R. Girshick, & A. Farhadi, “YOLOv4: Optimal Speed and Accuracy of Object Detection,” arXiv:1506.02640 [cs.CV], 2015.
[50] Liu, S., Qi, L., Qin, H., Shi, J., & Jia, J., “Path Aggregation Network for Instance Segmentation.” 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018.
[51] “ROS/Introduction – ROS Wiki,” Accessed on: April. 20th, 2021 [Online]. Available: http://wiki.ros.org/ROS/Introduction.
[52] “Robot Operating System.” Wikipedia, Wikimedia Foundation, 5 Apr. 2022, https://en.wikipedia.org/wiki/Robot_Operating_System.
[53] Tome, D., Russell, C., & Agapito, L., “Lifting from the Deep: Convolutional 3D Pose Estimation from a Single Image.” 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
[54] Stereo Labs, “zed-ros-wrapper” [Source code] Available: https://github.com/stereolabs/zed-ros-wrapper, 2021.
[55] G. Hidalgo, Z. Cao, T. Simon, S.E. Wei, Y. Raaj, H. Joo, Y. Sheikh., “Openpose,” Available: https://github.com/CMU-Perceptual-Computing-Lab/openpose.
[56] Massardi, J., M. Gravel, & Éric Beaudry., “Error-Tolerant Anytime Approach to Plan Recognition Using a Particle Filter”. Proceedings of the International Conference on Automated Planning and Scheduling, vol. 29, no. 1, May 2021, pp. 284-91, https://ojs.aaai.org/index.php/ICAPS/article/view/3490.
[57] “Terminal and Nonterminal Symbols.” Wikipedia, Wikimedia Foundation, 20 Apr. 2022, https://en.wikipedia.org/wiki/Terminal_and_nonterminal_symbols.
[58] Cappe, O., Godsill, S. J., & Moulines, E., “An Overview of Existing Methods and Recent Advances in Sequential Monte Carlo.” Proceedings of the IEEE, vol. 95, no. 5, 2007, pp. 899–924.
[59] Zhang, B., Liu, J., & Chen, H., “AMCL Based Map Fusion for Multi-Robot Slam with Heterogenous Sensors.” 2013 IEEE International Conference on Information and Automation (ICIA), 2013.
[60] Massardi, J., Gravel, M., & Beaudry, E., “Parc: A Plan and Activity Recognition Component for Assistive Robots.” 2020 IEEE International Conference on Robotics and Automation (ICRA), 2020.
[61] Oh, Jung H., & Heung-Jae Lee., “Object-Based Place Recognition for Mobile Robots Using Laplace's Rule of Succession.” Proceedings of the 2020 12th International Conference on Computer and Automation Engineering, 2020.
[62] Handa, A., Patraucean, V., Badrinarayanan, V., Stent, S., & Cipolla, R., “Understanding Realworld Indoor Scenes with Synthetic Data.” 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
[63] Hernandez, A. C., Gomez, C., Barber, R., & Martinez Mozos, O., “Object-Based Probabilistic Place Recognition for Indoor Human Environments.” 2018 International Conference on Control, Artificial Intelligence, Robotics & Optimization (ICCAIRO), 2018.
[64] Xiao, J., Hays, J., Ehinger, K. A., Oliva, A., & Torralba, A., “Sun Database: Large-Scale Scene Recognition from Abbey to Zoo.” 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2010.
[65] Song, Y.-F., Zhang, Z., Shan, C., & Wang, L., “Constructing Stronger and Faster Baselines for Skeleton-Based Action Recognition.” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022, pp. 1–1.
[66] He, K., Zhang, X., Ren, S., & Sun, J., “Deep Residual Learning for Image Recognition.” 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
[67] Leotta, F., Mecella, M., Sora, D., & Catarci, T., “Surveying Human Habit Modeling and Mining Techniques in Smart Spaces.” Future Internet, vol. 11, no. 1, 2019, p. 23.
[68] Shahroudy, A., Liu, J., Ng, T.-T., & Wang, G., “NTU RGB+D: A Large Scale Dataset for 3D Human Activity Analysis.” 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
[69] StereoLabs, “ZED 2,” Available: https://www.stereolabs.com/zed-2/.
[70] “Triangulation.” Wikipedia, Wikimedia Foundation, 18 July 2022, https://en.wikipedia.org/wiki/Triangulation.
[71] Byeon, Y.-H., Kim, D., Lee, J., & Kwak, K.-C., “Ensemble Three-Stream RGB-s Deep Neural Network for Human Behavior Recognition Under Intelligent Home Service Robot Environments.” IEEE Access, vol. 9, 2021, pp. 73240–73250.
[72] B. Zhou, A. Lapedriza, A. Khosla, A. Oliva, & A. Torralba, “Places: A 10 Million Image Database for Scene Recognition,” in IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 40, no. 6, pp. 1452-1464, 1 June 2018.
[73] Lee, Jaeryoung, & Nicholas Melo., “Habit Representation Based on Activity Recognition.” Sensors (Basel, Switzerland) vol. 20,7 1928. 30 Mar. 2020.
-
dc.identifier.urihttp://tdr.lib.ntu.edu.tw/jspui/handle/123456789/87342-
dc.description.abstract近年來技術的快速發展與提升,為機器人領域開闢了多項新的領域,許多研究大量投入居家機器人的應用及解決其可能遇到的問題,多年來,研究人員與專業人士將注意力轉向提升機器人的智慧能力,此提升不僅能提高程序的效率,更能使機器人更人性化,而居家機器人要有智慧就必須擁有理解周邊環境的能力與辨識人類行為和背後意義的能力,此外給予必要與適當的回饋與反應。
雖然在這個領域已經擁有非常多的研究,但還是存在許多須解決的問題和對人類行為理解上的挑戰。首先,對於一個居家型機器人而言,要具備有理解人類活動上的能力是至關重要的,因此,機器人須具備一個高準確度的人類活動辨識系統,除此之外,由於人類的活動與執行的地點有高度相關性,位置訊息被認為是有助於提高人類活動辨識系統的表現,因此,機器人也必須具備識別地點的能力,然而,當機器人感知影像中存在多個地點時,要獲得一個穩定且準確的位置估計技術是不容易也具挑戰性的。其次,為了理解一系列所觀察到的活動,計畫辨識扮演著至關重要的腳色,此外人類的習慣是影響序列中活動順序的重要因素,並能提高與改變計畫辨識的結果。最後,機器人與人類的互動是應用中最重要的部分,基於對於人類活動與目標的理解,機器人有必要運用這些知識和人類進行互動,並潛在的改變人類的行為與決策而使系統更有意義在。
本篇研究中,我們考慮了一個整合兩種不同估計地點模型的方法來獲得用於人類活動辨識系統的位置訊息,其中一個模型名為ResNet50-Place365是透過處理單一影像來估計地點,另一個是我們創建的模型,此模型利用人與物體之間的距離來決定人所在的位置。處此之外,我們提出一個基於適應性圖卷積及活動地點的人類活動辨識系統名為AL-GCN,該系統透過輸入三維人體骨架和估計的地點來預測居家環境中人類的行為,以實現更好的準確性與強健性。我們還設計了一個計劃辨識系統來預測下一個活動、目標與計畫,為了提高計畫辨識的性能,我們透過創建一個包含人類習慣的知識庫讓系統可以將人類習慣作為辨識的資訊,該知識庫包含存儲各種活動序列的plan library、用於預測下一個活動的Loc-NextAct Tensor與用於預測目標的Loc-Objective Tensor,此外該知識庫也能夠獲得新訊息並做適當的更新來改進預測的結果與適應不同的使用者。最後,偵測到的活動與預測將作為機器人響應模組的輸入來向用戶提供建議、警告與提醒。最終我們將系統部屬到我們自己的機器人上,以執行現實世界中會發生的場景。
在實驗中,我們進行了兩種不同類型的評估,一個是對數據集做的評估,另一個則是在真實世界中對我們的機器人所做的評估。在數據集的評估中,我們所提出將兩種不同位置估計融合的方法可獲得92.83%的準確率,而這個準確率都比僅使用其中任何一個都更準確,此外我們的AL-GCN模型透過結合位置訊息與人體骨架在Cross-subject的評估上可以達到94.33%的準確率,而我們的計畫辨識系統透過考慮位置訊息與更新知識庫來改進預測,並可得知系統在活動序列的前幾次觀測就可以獲得較高準確的預測。在真實世界中,我們的位置估計方法在估計客廳上可以達到98%的準確率,至於提出的AL-GCN模型加上位置訊息後準確率提高了10%到20%,最後,在執行不同的計劃時,我們提出的計畫辨識系統所作的預測會隨著知識庫的更新而有顯著的提高。
在未來研究上,我們的系統可以與Re-ID做結合使機器人能分辨不同的人物,並進一步使用系統所儲存的個別個人習慣進行預測,除此之外,機器人響應可以進一步考慮習慣時間上的資訊,並提供一個在時間上更精確地回應。
zh_TW
dc.description.abstractThe rapid development and growth of technology has widened up and opened new frontiers of the field of robotics. Many researches have devoted in investigating indoor robotic applications and the problems they may encounter. Throughout the years, researchers and professionals have turn their focus on producing robots that are becoming more and more intelligent. The increase of intelligence in robots not only allow the program to be more efficient, but also allow the robots to be more humanized. For household robots to be more intelligent, they require the ability to understand the surrounding environment and human behavior. Moreover, give necessary and suitable feedback and responses when interacting with a human.
In this field of research, there are still much to investigate and many problems to resolve. First, a household robot that possess the ability to understand human activity is crucial. Thus, a human activity recognition (HAR) system that has a high recognition accuracy is desired. Moreover, due to the high relationship between indoor activities and location, the location information is considered helpful for improving the HAR. Hence, the robot also needs to possess the ability to recognize location. However, obtaining a stable and accurate location estimating technique is challenging when the robot perceives multiple locations in an image. Second, to understand a sequence of perceived activities, plan recognition plays a vital role. Moreover, human habit is an essential element that affects the order of activities in the sequences which can enhance the plan recognition performance. Last but not least, the interaction between robot and user is a major part in robot application. Upon the knowledges from the understanding of human activity and plan recognition, the robot is necessary to utilize these knowledges to interact with humans and potentially impact human decision which makes the system more meaningful.
In this study, we consider a location estimation method that integrates two different models to obtain the location information that is used for HAR. A model called ResNet50-Place365 is utilized to estimate location by processing image data. The other model is created by us called the location estimator, which utilizes the distance between human and objects to determine which location the human is in. Moreover, we propose a human activity recognition system called activity-location graph convolutional neural network (AL-GCN) based on adaptive graph convolutional network that incorporates location information to understand human behaviors. This model predicts the human activity in a household environment by taking the three-dimensional human skeleton and the estimated location as input to achieve better prediction accuracy and robustness. We also propose a plan recognition system that gives various predictions, such as next activity, objective, and plan. In order to enhance the performance of plan recognition, we take the human habit into account by creating a knowledge base that consists of a plan library, which stores various sequence of activities, a Loc-NextAct Tensor, which is used for predicting the next activity, and a Loc-Objective Tensor, which is used for predicting the objective. Moreover, the knowledge base can be updated as new information are obtained to adapt to different users and improve prediction results. Furthermore, a response module that takes the activity and predictions results from HAR and plan recognition as inputs is created to give advices, warnings, or reminders to the user. Finally, our system is physically deployed onto our own home-made robot for executing real-world scenarios.
In our experiment, two types of evaluations are conducted. One is the evaluation on the datasets, and the other is the evaluation on our physical robot in real-world scenarios. In the dataset evaluation, our proposed method that fuses ResNet50-Place365 and location estimator receives an accuracy of 92.83%, which is more accurate than only using either one of them. By incorporating the location information, our proposed AL-GCN HAR model achieves a 94.33% accuracy on the cross-subject evaluation. Moreover, the predictions of our plan recognition system are improved by updating the knowledge base and considering the location information. As a result, the predictions can achieve high accuracy in the first few observations. In the real-world experiment, our location estimation model achieves a 98% accuracy in estimating the living room. As for our AL-GCN Model, with the location information added, the accuracy of different activities gain improvement ranging from 10% to 20%. Finally, our plan recognition results in real-world show that with the updated knowledge base, the accuracy prediction increases significantly.
In Future work, our system can integrate with Re-ID to allow the robot to recognize multiple people and further apply their personal stored habit for predictions. Moreover, robot responses can take the habit time into consideration and provide a more precise response.
en
dc.description.provenanceSubmitted by admin ntu (admin@lib.ntu.edu.tw) on 2023-05-18T17:10:33Z
No. of bitstreams: 0
en
dc.description.provenanceMade available in DSpace on 2023-05-18T17:10:33Z (GMT). No. of bitstreams: 0en
dc.description.tableofcontents論文口試委員會審定書 I
致謝 II
摘要 IV
ABSTRACT VI
TABLE OF CONTENTS IX
LIST OF FIGURES XIII
LIST OF TABLES XVI
Chapter 1 Introduction 1
1.1 Motivation 1
1.2 Research objectives 3
1.3 Related Works 4
1.3.1 Human Activity Recognition (HAR) 5
1.3.2 Plan Recognition 12
1.4 Contribution 17
1.5 Thesis Overview 18
Chapter 2 Preliminary 20
2.1 Graph Convolutional Network (GCN) 20
2.2 Spatial-Temporal Graph Convolutional Network (ST-GCN) 23
2.3 Adaptive GCN 24
2.4 YOLO v4 Object Detection 25
2.5 Robot Operating System (ROS) 27
2.6 3D Human Pose Estimation 29
2.7 Plan Library 30
2.8 Particle Filter 32
Chapter 3 Methodology 34
3.1 System Overview 34
3.2 Overview of Functionalities 35
3.3 Location Estimation 36
3.3.1 Location Definition 36
3.3.2 Representative Objects and Affordance Objects 37
3.3.3 Estimating Location 39
3.4 Recognition Repositioning 42
3.4.1 Human Heading Detection 43
3.4.2 Perception Point: Distance and Section Selection 43
3.4.3 Recognition Reposition Flow 45
3.5 Human Activity Recognition (HAR) 46
3.5.1 Human Activity Recognition with Adaptive GCN 46
3.5.2 Human Activity Recognition Model (AL-GCN) 47
3.6 Plan Recognition 49
3.6.1 Plan Recognition System flow 50
3.6.2 Plan Library Definition and Construction 50
3.6.3 Human Habit Incorporation 52
3.6.4 Plan Predictions 56
3.6.5 Knowledge Update 59
3.7 Robot Response Module 62
3.7.1 Different Response for Different Activities 62
Chapter 4 Experiment 64
4.1 Experiment Setup 64
4.1.1 Human Activity Recognition 64
4.1.2 Plan Recognition 66
4.1.3 Robot setup: Oreo 68
4.1.4 Sensor setup: ZED 2 Camera 68
4.1.5 Environmental setup: Yonglin Lab 70
4.2 Location Estimation 71
4.2.1 Experiment and Evaluation on ETRI Dataset 71
4.3 Human Activity Recognition (AL-GCN Model) 73
4.3.1 Experiment and Evaluation on ETRI Dataset 73
4.4 Plan Recognition 82
4.4.1 Next Activity Prediction 82
4.4.2 Objective Prediction 83
4.4.3 Plan Prediction 85
4.5 Overall System Result in Real World Robot 87
4.5.1 System Flow Performed by OREO Robot in Experiment 88
4.5.2 Location Estimator on Oreo Robot 89
4.5.3 HAR Systen on Oreo Robot 91
4.5.4 Plan Recognition System on Oreo Robot 95
Chapter 5 Conclusion and Future Works 97
REFERENCES 99
Appendix I 105
Appendix II 106
Appendix III 107
Appendix IV 121
-
dc.language.isoen-
dc.subject居家機器人zh_TW
dc.subject機器人響應zh_TW
dc.subject人類習慣zh_TW
dc.subject人類活動辨識zh_TW
dc.subject計畫辨識zh_TW
dc.subject地點估計zh_TW
dc.subjectrobot responseen
dc.subjecthousehold roboten
dc.subjecthuman habiten
dc.subjectplan recognitionen
dc.subjectHuman activity recognitionen
dc.subjectlocation estimationen
dc.title居家型機器人運用位置訊息進行人類活動與習慣理解zh_TW
dc.titleHousehold Robot Utilizing Location Information for Human Activity and Habit Understandingen
dc.typeThesis-
dc.date.schoolyear111-1-
dc.description.degree碩士-
dc.contributor.oralexamcommittee陳永耀;楊谷洋;張文中;曾士桓zh_TW
dc.contributor.oralexamcommitteeYung-Yaw Chen;Kuu-Young Young;Wen-Chung Chang;Shih-Huan Tsengen
dc.subject.keyword人類活動辨識,計畫辨識,地點估計,居家機器人,人類習慣,機器人響應,zh_TW
dc.subject.keywordHuman activity recognition,plan recognition,location estimation,household robot,human habit,robot response,en
dc.relation.page157-
dc.identifier.doi10.6342/NTU202300261-
dc.rights.note同意授權(全球公開)-
dc.date.accepted2023-02-08-
dc.contributor.author-college電機資訊學院-
dc.contributor.author-dept電機工程學系-
dc.date.embargo-lift2026-03-01-
顯示於系所單位:電機工程學系

文件中的檔案:
檔案 大小格式 
ntu-111-1.pdf9.56 MBAdobe PDF檢視/開啟
顯示文件簡單紀錄


系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。

社群連結
聯絡資訊
10617臺北市大安區羅斯福路四段1號
No.1 Sec.4, Roosevelt Rd., Taipei, Taiwan, R.O.C. 106
Tel: (02)33662353
Email: ntuetds@ntu.edu.tw
意見箱
相關連結
館藏目錄
國內圖書館整合查詢 MetaCat
臺大學術典藏 NTU Scholars
臺大圖書館數位典藏館
本站聲明
© NTU Library All Rights Reserved