Skip navigation

DSpace

機構典藏 DSpace 系統致力於保存各式數位資料(如:文字、圖片、PDF)並使其易於取用。

點此認識 DSpace
DSpace logo
English
中文
  • 瀏覽論文
    • 校院系所
    • 出版年
    • 作者
    • 標題
    • 關鍵字
    • 指導教授
  • 搜尋 TDR
  • 授權 Q&A
    • 我的頁面
    • 接受 E-mail 通知
    • 編輯個人資料
  1. NTU Theses and Dissertations Repository
  2. 電機資訊學院
  3. 電機工程學系
請用此 Handle URI 來引用此文件: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/86373
完整後設資料紀錄
DC 欄位值語言
dc.contributor.advisor傅立成(Li-Chen Fu)
dc.contributor.authorPei-Han Huangen
dc.contributor.author黃珮涵zh_TW
dc.date.accessioned2023-03-19T23:52:01Z-
dc.date.copyright2022-08-31
dc.date.issued2022
dc.date.submitted2022-08-24
dc.identifier.citation[1] L. F. Jacobs and F. Schenk, 'Unpacking the cognitive map: the parallel map theory of hippocampal function,' Psychological review, vol. 110, no. 2, pp. 285-315, 2003. [2] E. C. Tolman, 'Cognitive maps in rats and men,' Psychological Review, vol. 55, no. 4, pp. 189-208, 1948. [3] R. A. Epstein, E. Z. Patai, J. B. Julian, and H. J. Spiers, 'The cognitive map in humans: spatial navigation and beyond,' Nature Neuroscience, vol. 20, no. 11, pp. 1504-1513, 2017. [4] H. R. Evensmoen, H. Lehn, J. Xu, M. P. Witter, L. Nadel, and A. K. Håberg, 'The anterior hippocampus supports a coarse, global environmental representation and the posterior hippocampus supports fine-grained, local environmental representations,' J Cogn Neurosci, vol. 25, no. 11, pp. 1908-25, 2013. [5] K. B. Iva, B. Buddhika, D. O. Jason, M. Vincent, R. Jessica, L. Zhong-Xu, G. Cheryl, R. S. Rosenbaum, W. Gordon, D. B. Morgan, and M. Morris, 'Multiple Scales of Representation along the Hippocampal Anteroposterior Axis in Humans,' Current Biology, vol. 28, no. 13, pp. 2129-2135, 2018. [6] M. Peer, Y. Ron, R. Monsa, and S. Arzy, 'Processing of different spatial scales in the human brain,' eLife, vol. 8, pp. e47492, 2019. [7] J. L. S. Bellmund, P. Gärdenfors, E. I. Moser, and C. F. Doeller, 'Navigating cognition: Spatial codes for human thinking,' Science, vol. 362, no. 6415, 2018. [8] T. E. J. Behrens, T. H. Muller, J. C. R. Whittington, S. Mark, A. B. Baram, K. L. Stachenfeld, and Z. Kurth-Nelson, 'What Is a Cognitive Map? Organizing Knowledge for Flexible Behavior,' Neuron, vol. 100, no. 2, pp. 490-509, Oct. 2018. [9] M. Schafer and D. Schiller, 'Navigating Social Space,' Neuron, vol. 100, no. 2, pp. 476-489, 2018. [10] E. A. Maguire, 'The retrosplenial contribution to human navigation: A review of lesion and neuroimaging findings,' Scandinavian Journal of Psychology, vol. 42, no. 3, pp. 225-238, 2001. [11] R. A. Epstein, 'Parahippocampal and retrosplenial contributions to human spatial navigation,' Trends Cogn Sci, vol. 12, no. 10, pp. 388-96, 2008. [12] J. O'keefe and L. Nadel, 'Précis of O'Keefe & Nadel's The hippocampus as a cognitive map,' Behavioral and Brain Sciences, vol. 2, no. 4, pp. 487-494, 1979. [13] D. R. Montello,'A New Framework for Understanding the Acquisition of Spatial Knowledge in Large-Scale Environments,' in Spatial and Temporal Reasoning in Geographic Information Systems, UK: Oxford University Press, 1998. [14] R. G. Golledge,'Human wayfinding and cognitive maps,' in Wayfinding behavior : cognitive mapping and other spatial processes, USA: Johns Hopkins University Press, 1999. [15] T. Meilinger, M. Strickrodt, and H. H. Bülthoff, 'Spatial Survey Estimation Is Incremental and Relies on Directed Memory Structures,' in Spatial Cognition XI, 2018, pp. 27-42. [16] B. Kuipers, D. G. Tecuci, and B. J. Stankiewicz, 'The skeleton in the cognitive map: A computational and empirical exploration,' Environment and Behavior, vol. 35, no. 1, pp. 81-106, 2003. [17] A.-H. Javadi, B. Emo, L. R. Howard, F. E. Zisch, Y. Yu, R. Knight, J. Pinelo Silva, and H. J. Spiers, 'Hippocampal and prefrontal processing of network topology to simulate the future,' Nature Communications, vol. 8, no. 1, pp. 14652, 2017. [18] G. Janzen and M. van Turennout, 'Selective neural representation of objects relevant for navigation,' Nat Neurosci, vol. 7, no. 6, pp. 673-677, 2004. [19] S. A. Marchette, L. K. Vass, J. Ryan, and R. A. Epstein, 'Outside looking in: Landmark generalization in the human navigational system,' The Journal of Neuroscience, vol. 35, no. 44, pp. 14896-14908, 2015. [20] T. Meilinger, 'The Network of Reference Frames Theory: A Synthesis of Graphs and Cognitive Maps,' in Spatial Cognition VI. Learning, Reasoning, and Talking about Space, 2008, pp. 344-360. [21] W. H. Warren, 'Non-Euclidean navigation,' J Exp Biol, vol. 222, no. Pt Suppl 1, 2019. [22] M. Peer, I. K. Brunec, N. S. Newcombe, and R. A. Epstein, 'Structuring Knowledge with Cognitive Maps and Cognitive Graphs,' Trends in Cognitive Sciences, vol. 25, no. 1, pp. 37-54, 2021. [23] D. S. Chaplot, D. Gandhi, A. Gupta, and R. Salakhutdinov, 'Object Goal Navigation Using Goal-Oriented Semantic Exploration,' in 34th International Conference on Neural Information Processing Systems (NIPS'20), 2020, pp. 4247–4258. [24] H. Carrillo, I. Reid, and J. A. Castellanos, 'On the comparison of uncertainty criteria for active SLAM,' in IEEE International Conference on Robotics and Automation (ICRA), 2012, pp. 2080-2087. [25] L. Shen, Z. Lin, and Q. Huang, 'Relay Backpropagation for Effective Learning of Deep Convolutional Neural Networks,' in European Conference on Computer Vision (ECCV), 2016, pp. 467-482. [26] J. Ryu, M.-H. Yang, and J. Lim, 'DFT-based Transformation Invariant Pooling Layer for Visual Classification,' in European Conference on Computer Vision (ECCV), 2018, pp. 89-104. [27] L. Berrada, A. Zisserman, and M. P. Kumar, 'Deep Frank-Wolfe for neural network optimization,' arXiv:1811.07591, 2018. [28] M. George, M. Dixit, G. Zogg, and N. Vasconcelos, 'Semantic Clustering for Robust Fine-Grained Scene Recognition,' in European Conference on Computer Vision (ECCV), 2016, pp. 783-798. [29] J. H. Bappy, S. Paul, and A. K. Roy-Chowdhury, 'Online Adaptation for Joint Scene and Object Classification,' in European Conference on Computer Vision (ECCV), 2016, pp. 227-243. [30] L. Torresani, M. Szummer, and A. Fitzgibbon, 'Efficient Object Category Recognition Using Classemes,' in European Conference on Computer Vision (ECCV), 2010, pp. 776-789. [31] A. Bergamo and L. Torresani, 'Classemes and Other Classifier-Based Features for Efficient Object Categorization,' IEEE Trans Pattern Anal Mach Intell, vol. 36, no. 10, pp. 1988-2001, 2014. [32] L.-J. Li, H. Su, Y. Lim, and L. Fei-Fei, 'Object Bank: An Object-Level Image Representation for High-Level Visual Recognition,' International Journal of Computer Vision, vol. 107, no. 1, pp. 20-39, 2014. [33] Z. Wang, L. Wang, Y. Wang, B. Zhang, and Y. Qiao, 'Weakly Supervised PatchNets: Describing and Aggregating Local Patches for Scene Recognition,' IEEE Transactions on Image Processing, vol. 26, no. 4, pp. 2028-2041, 2017. [34] X. Cheng, J. Lu, J. Feng, B. Yuan, and J. Zhou, 'Scene recognition with objectness,' Pattern Recognition, vol. 74, pp. 474-487, 2018. [35] S. Jiang, G. Chen, X. Song, and L. Liu, 'Deep Patch Representations with Shared Codebook for Scene Classification,' ACM Trans. Multimedia Comput. Commun. Appl., vol. 15, no. 1s, pp. 1-17, 2019. [36] L. Wang, S. Guo, W. Huang, Y. Xiong, and Y. Qiao, 'Knowledge guided disambiguation for large-scale scene classification with multi-resolution CNNs,' IEEE Transactions on Image Processing, vol. 26, no. 4, pp. 2055-2068, 2017. [37] X. Sun, L. Zhang, Z. Wang, J. Chang, Y. Yao, P. Li, and R. Zimmermann, 'Scene categorization using deeply learned gaze shifting kernel,' IEEE transactions on cybernetics, vol. 49, no. 6, pp. 2156-2167, 2018. [38] N. Sun, W. Li, J. Liu, G. Han, and C. Wu, 'Fusing object semantics and deep appearance features for scene recognition,' IEEE Transactions on Circuits and Systems for Video Technology, vol. 29, no. 6, pp. 1715-1728, 2018. [39] J. Redmon and A. Farhadi, 'YOLO9000: better, faster, stronger,' in IEEE conference on computer vision and pattern recognition (CVPR), 2017, pp. 7263-7271. [40] H. Seong, J. Hyun, H. Chang, S. Lee, S. Woo, and E. Kim, 'Scene Recognition via Object-to-Scene Class Conversion: End-to-End Training,' in International Joint Conference on Neural Networks (IJCNN), 2019, pp. 1-6. [41] H. Seong, J. Hyun, and E. Kim, 'FOSNet: An End-to-End Trainable Deep Neural Network for Scene Recognition,' IEEE Access, vol. 8, pp. 82066-82077, 2020. [42] J. Crespo, J. C. Castillo, O. M. Mozos, and R. Barber, 'Semantic Information for Robot Navigation: A Survey,' Applied Sciences, vol. 10, no. 2, pp. 497, 2020. [43] H. Zender, O. M. Mozos, P. Jensfelt, G.-J. Kruijff, and W. Burgard, 'Conceptual spatial representations for indoor mobile robots,' Robotics and Autonomous Systems, vol. 56, no. 6, pp. 493-502, 2008. [44] C. Nieto-Granda, J. G. Rogers, A. J. Trevor, and H. I. Christensen, 'Semantic map partitioning in indoor environments using regional analysis,' in IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2010, pp. 1451-1456. [45] R. Barber, J. Crespo, C. Gómez, A. C. Hernámdez, and M. Galli, 'Mobile robot navigation in indoor environments: Geometric, topological, and semantic navigation,' in Applications of Mobile Robots: IntechOpen, 2018. [46] I. Kostavelis, K. Charalampous, A. Gasteratos, and J. K. Tsotsos, 'Robot navigation via spatial and temporal coherent semantic maps,' Engineering Applications of Artificial Intelligence, vol. 48, pp. 173-187, 2016. [47] D. Joho and W. Burgard, 'Searching for objects: Combining multiple cues to object locations using a maximum entropy model,' in IEEE International Conference on Robotics and Automation (ICRA), 2010, pp. 723-728. [48] D. S. Chaplot, H. Jiang, S. Gupta, and A. Gupta, 'Semantic Curiosity for Active Visual Learning,' in European Conference on Computer Vision (ECCV), 2020, pp. 309-326. [49] K. Bollacker, C. Evans, P. Paritosh, T. Sturge, and J. Taylor, 'Freebase: A Collaboratively Created Graph Database for Structuring Human Knowledge,' in International Conference on Management of Data (SIGMOD), 2008, pp. 1247-1250. [50] J. Lehmann, R. Isele, M. Jakob, A. Jentzsch, D. Kontokostas, P. N. Mendes, S. Hellmann, M. Morsey, P. van Kleef, S. Auer, and C. Bizer, 'A Large-scale, Multilingual Knowledge Base Extracted from Wikipedia,' Semantic Web, vol. 6, no. 2, pp. 167-195, 2015. [51] T. Thomas Pellissier, V. Denny, S. Sebastian, S. Thomas, and P. Lydia, 'From Freebase to Wikidata: The Great Migration,' in 25th International Conference on World Wide Web (WWW), 2016, pp. 1419-1428. [52] R. Speer, J. Chin, and C. Havasi, 'ConceptNet 5.5: an open multilingual graph of general knowledge,' in Thirty-First AAAI Conference on Artificial Intelligence, 2017, pp. 4444-4451. [53] J. Lei, W. Yujing, S. Botian, Z. Dawei, W. Zhongyuan, and Y. Jun, 'Microsoft Concept Graph: Mining Semantic Concepts for Short Text Understanding,' Data Intelligence, vol. 1, no. 3, pp. 238-270, 2019. [54] Y. Qin, H. Cao, and L. Xue, 'Research and Application of Knowledge Graph in Teaching: Take the database course as an example,' Journal of Physics: Conference Series, vol. 1607, pp. 012127, 2020. [55] Y. Lan, G. He, J. Jiang, J. Jiang, W. X. Zhao, and J.-R. Wen, 'A survey on complex knowledge base question answering: Methods, challenges and solutions,' arXiv:2105.11644, 2021. [56] M. Quigley, K. Conley, B. Gerkey, J. Faust, T. Foote, J. Leibs, R. Wheeler, and A. Y. Ng, 'ROS: an open-source Robot Operating System,' in ICRA workshop on open source software, vol. 3, no. 3.2, pp. 5, 2009. [57] L. F. Jacobs, 'From chemotaxis to the cognitive map: The function of olfaction,' National Academy of Sciences, pp. 10693-10700, 2012. [58] D. Fernández-Fernández,'Establishment, validation and pharmacological manipulation of short and long-term potentiation in murine hippocampal slices,' Oct. 2015. [59] E. T. Rolls, 'The storage and recall of memories in the hippocampo-cortical system,' Cell Tissue Res., vol. 373, no. 3, pp. 577-604, 2018. [60] B. Yamauchi, 'A frontier-based approach for autonomous exploration,' in IEEE International Symposium on Computational Intelligence in Robotics and Automation (CIRA), 1997, pp. 146-151. [61] Z. Qilong and R. Pless, 'Extrinsic calibration of a camera and laser range finder (improves camera calibration),' in IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2004, pp. 2301-2306. [62] L. McInnes, o. Healy, and S. Astels. 'How HDBSCAN Works.' https://hdbscan.readthedocs.io/en/latest/how_hdbscan_works.html (accessed May 27, 2022). [63] S. I. Gass and M. C. Fu, 'Prim’s Algorithm,' in Encyclopedia of Operations Research and Management Science. Boston, MA: Springer US, 2013, p. 1160. [64] P. E. Hart, N. J. Nilsson, and B. Raphael, 'A formal basis for the heuristic determination of minimum cost paths,' IEEE transactions on Systems Science and Cybernetics, vol. 4, no. 2, pp. 100-107, 1968. [65] E. W. Dijkstra, 'A note on two problems in connexion with graphs,' in Edsger Wybe Dijkstra: His Life, Work, and Legacy, 2022, pp. 287-290. [66] D. Fox, W. Burgard, and S. Thrun, 'The dynamic window approach to collision avoidance,' IEEE Robotics & Automation Magazine, vol. 4, no. 1, pp. 23-33, 1997. [67] E. Marder-Eppstein. 'dwa_local_planner.' http://wiki.ros.org/dwa_local_planner (accessed July 2, 2022). [68] B. Zhou, H. Zhao, X. Puig Fernandez, T. Xiao, S. Fidler, A. Barriuso, and A. Torralba, 'Semantic Understanding of Scenes Through the ADE20K Dataset,' International Journal of Computer Vision, pp. 302-321, 2019. [69] S. Woo, J. Park, J.-Y. Lee, and I. S. Kweon, 'CBAM: Convolutional Block Attention Module,' in European Conference on Computer Vision (ECCV), vol. 11211, 2018, pp. 3-19. [70] C. Szegedy, L. Wei, J. Yangqing, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich, 'Going deeper with convolutions,' in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015, pp. 1-9. [71] A. Elfes, 'Using occupancy grids for mobile robot perception and navigation,' Computer, vol. 22, no. 6, pp. 46-57, 1989. [72] N. Sünderhauf, F. Dayoub, S. McMahon, B. Talbot, R. Schulz, P. Corke, G. Wyeth, B. Upcroft, and M. Milford, 'Place categorization and semantic mapping on a mobile robot,' in IEEE International Conference on Robotics and Automation (ICRA), 2016, pp. 5729-5736. [73] M. Honnibal and I. Montani, 'spaCy 2: Natural language understanding with Bloom embeddings, convolutional neural networks and incremental parsing,' 2017. [74] N. Reimers and I. Gurevych, 'Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks,' arXiv:1908.10084, 2019. [75] A. Zhang. 'SpeechRecognition.' https://pypi.org/project/SpeechRecognition/ (accessed July 3, 2022). [76] S. Han. 'googletrans.' https://pypi.org/project/googletrans/ (accessed July 3, 2022). [77] ROBOTIS. 'turtlebot3_simulations.' https://github.com/ROBOTIS-GIT/turtlebot3_simulations (accessed July 4, 2022). [78] A. Robotics. 'aws-robomaker-small-house-world.' https://github.com/aws-robotics/aws-robomaker-small-house-world (accessed July 4, 2022). [79] A. Topiwala, P. Inani, and A. Kathpal, 'Frontier Based Exploration for Autonomous Robot,' arXiv:1806.03581, 2018. [80] A. Quattoni and A. Torralba, 'Recognizing indoor scenes,' in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2009, pp. 413-420. [81] B. Zhou, H. Zhao, X. Puig, S. Fidler, A. Barriuso, and A. Torralba, 'Scene Parsing through ADE20K Dataset,' in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 5122-5130. [82] B. Zhou, A. Lapedriza, A. Khosla, A. Oliva, and A. Torralba, 'Places: A 10 Million Image Database for Scene Recognition,' IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 40, no. 6, pp. 1452-1464, 2018. [83] K. He, X. Zhang, S. Ren, and J. Sun, 'Deep residual learning for image recognition,' in IEEE conference on computer vision and pattern recognition (CVPR), 2016, pp. 770-778. [84] A. López-Cifuentes, M. E.-V. J. Bescós, and Á. García-Martín, 'Semantic-aware scene recognition,' Pattern Recognition, vol. 102, pp. 107256, 2020. [85] A. Grover and J. Leskovec, 'node2vec: Scalable feature learning for networks,' in Proceedings of the 22nd ACM SIGKDD international conference on Knowledge discovery and data mining, 2016, pp. 855-864. [86] C.-H. Chen, M.-F. Shiu, and S.-H. Chen, 'Use Learnable Knowledge Graph in Dialogue System for Visually Impaired Macro Navigation,' Applied Sciences, vol. 11, no. 13, pp. 6057, 2021.
dc.identifier.urihttp://tdr.lib.ntu.edu.tw/jspui/handle/123456789/86373-
dc.description.abstract全球正面臨著少子化及人口高齡化的趨勢,勞動力缺口及高齡長照等替服務型機器人帶來直接的需求,進而衍伸了許多如居家協助與照護等問題,而為了能夠使機器人能夠適應不同場合以提供各式的服務,機器人的自主環境認知、推理決策與導航能力尤為重要,此外,服務型機器人亦應能夠理解人類語言,才能在人機互動中做出適當的移動決策,另一方面,此類型的機器人通常配載有限的運算資源與感測器,在整體系統設計架構中必須選擇能夠同時兼顧準確性與運算效能等問題。 在本篇研究當中,我們提出了一個基於自主環境認知地圖的推理導航架構,並使用最常見的2D laser 與 RGBD 相機做為環境辨識的感測器,讓機器人進入一個新環境時,可以利用自主探索與環境辨識的能力,認知室內環境的場景配置,而且在自主探索的過程中,必須避免碰撞問題的發生,除了使用2D laser掃描平面環境外,在機器人正前方也使用深度相機點雲的融合,以辨識障礙物的真實邊界,並使用階層式的路徑規劃演算法,將點雲融合的訊息考慮進本地規劃器決策中,讓機器人也能夠閃避桌、椅等中空的家俱,在探索的路徑上,我們連續建構多個拓譜點,並在探索結束後利用場景語義網格圖,賦予拓譜點最具代表性的場景語義標籤。 於此同時,我們使用自然語言的數據集,建構在場景中常見的物體資訊與場景用途的語義知識圖譜,在機器人完成自主環境認知後,機器人就可以透過與使用者的對話,利用句子嵌入向量比對人類意圖與知識圖譜實體中的關聯性,推論出目標場景在認知地圖中的位置,最後使用帶有場景語義訊息的拓譜圖與A*路徑規劃演算法,規劃出到達目標場景的最短拓譜節點路徑提供服務。zh_TW
dc.description.abstractThe world is facing the trend of low birth rates and the aging of the population. The shortage of caregivers and more and more long-term care for the elderly has brought direct demand for service-oriented robots, which has further exacerbated many problems such as home assistance and care. To enable robots to adapt to be able to different environments and to provide various services, the ability of autonomous environment cognition, reasoning, decision-making, and navigation of robots are particularly important. In addition, a service-oriented robot should also be able to understand what human speaks so that it can make appropriate decisions during human-robot interaction. On the other hand, this type of robot usually only carries limited computing resources and common sensors. In the design of the overall system architecture, the robot must be able to take both accuracy and computing efficiency of the inferencing into account. In this research work, we propose a user’s intent driven navigation architecture based on the cognition of the indoor scene’s map and use the RGB-D camera for scene recognition. When the robot enters a new indoor environment, it can autonomously explore the environment and recognize the scene configuration while ensuring collision with obstacles can be avoided. During the process of collision avoidance, the 2D laser sensors not only tries to scan to detect obstacles on the fixed plane slightly above the ground, but also integrate the deep camera point cloud in front of the robot to reconstruct the real boundaries of the objects. Then, we design a hierarchical path planning framework such that the information of the depth point cloud will be utilized appropriately in the local planner enabling the robot to safely dodge the table, chair, and any other hollow furnitures featured with hollow bottom and thin legs. On the exploration trajectories, we continuously construct multiple topological nodes and established a topological map. After the exploration, we leverage the grid map and scene recognition to give the most representative scene labels to those nodes on the topological map. Besides collecting the geometric information and associated semantic meanings of the environment, we also use the dataset of natural language to construct a semantic knowledge graph that aggregates object information and major functions over different scenes. After the robot becomes sufficiently cognitive of the physical environment, it can infer the target scene and associated location to head to by investigating the correlation between the human intent and entity on the knowledge graph. Finally, given the inferred target scene, the robot maps out the relevant semantic topological nodes on the cognitive map and then applies the A* path planning algorithm to search for the shortest topological node path leading to the target scene so that the robot can accomplish the intended services.en
dc.description.provenanceMade available in DSpace on 2023-03-19T23:52:01Z (GMT). No. of bitstreams: 1
U0001-2308202217491500.pdf: 6900631 bytes, checksum: d120073fac7f4d26004cfda92e12ac06 (MD5)
Previous issue date: 2022
en
dc.description.tableofcontents口試委員會審定書 # 致謝 I 摘要 II ABSTRACT III TABLE OF CONTENTS V LIST OF FIGURES VIII LIST OF TABLES XI Chapter 1 Introduction 1 1.1 Background and Motivation 1 1.2 Research objectives 3 1.3 Contributions 4 1.4 Thesis Overview 5 Chapter 2 Background and Related Works 7 2.1 Introduction of Cognitive Map 8 2.1.1 Active SLAM 10 2.1.2 Scene Recognition 13 2.2 Target-driven Navigation 15 2.3 Knowledge Graph 19 2.4 ROS (Robot Operating System) 22 Chapter 3 Methodology 25 3.1 System Overview 25 3.2 Cognitive Map Construction 26 3.2.1 Parallel Map Theory 26 3.2.2 Component of Robot Cognitive Map 28 3.3 Exploration Module 28 3.3.1 Technical Basics 29 3.3.1.1 Introduction of Stereo Camera 30 3.3.1.2 Perception Module 31 3.3.1.3 Laser-Camera Calibration 32 3.3.2 Frontier Exploring Algorithm 34 3.3.3 Navigation Framework 39 3.3.4 Topological Node Construction 41 3.4 Scene Recognition Module 42 3.4.1 Object Branch 43 3.4.2 Image Branch 44 3.4.3 Fusion Module 45 3.4.4 Semantic Map 46 3.5 Intent-driven Navigation 49 3.5.1 Preliminary 49 3.5.1.1 Introduction of spaCy 49 3.5.1.2 Introduction of Sentence-BERT 51 3.5.2 Construction of Knowledge Graph 52 3.5.2.1 Data Cleaning 53 3.5.3 Scene Inference 56 3.5.3.1 spaCy Analysis 56 3.5.3.2 Reasoning Process 57 3.5.3.3 Topological Navigation 60 Chapter 4 Experiment 63 4.1 Exploring Module 63 4.2 Scene Recognition 67 4.2.1 Dataset Evaluation 67 4.2.2 Ablation study 72 4.2.2.1 Image Branch 72 4.2.2.2 Object Branch 73 4.2.2.3 Fusion Module 75 4.3 Knowledge Graph Evaluation 78 4.4 Real-world Experiment 82 Chapter 5 Conclusion and Future Works 91 REFERENCE 93 APPENDIX Ⅰ 99 APPENDIX Ⅱ 100 APPENDIX Ⅲ 101 APPENDIX Ⅳ 102
dc.language.isoen
dc.subject知識圖譜zh_TW
dc.subject語義認知地圖zh_TW
dc.subject環境辨識zh_TW
dc.subject主動定位與建圖zh_TW
dc.subjectKnowledge Graphen
dc.subjectActive SLAMen
dc.subjectScene Recognitionen
dc.subjectSemantic Cognitive Mapen
dc.title基於場景語義認知的意圖驅動導航居家服務型機器人zh_TW
dc.titleUser Intent-driven Navigation of Home Service Robot based on Semantic Scene Cognitionen
dc.typeThesis
dc.date.schoolyear110-2
dc.description.degree碩士
dc.contributor.oralexamcommittee宋開泰(Kai-Tai Song),張文中(Wen-Chung Chang),林沛群(Pei-Chun Lin),曾士桓(Shih-Huan Tseng)
dc.subject.keyword主動定位與建圖,環境辨識,語義認知地圖,知識圖譜,zh_TW
dc.subject.keywordActive SLAM,Scene Recognition,Semantic Cognitive Map,Knowledge Graph,en
dc.relation.page102
dc.identifier.doi10.6342/NTU202202718
dc.rights.note同意授權(全球公開)
dc.date.accepted2022-08-24
dc.contributor.author-college電機資訊學院zh_TW
dc.contributor.author-dept電機工程學研究所zh_TW
dc.date.embargo-lift2025-08-25-
顯示於系所單位:電機工程學系

文件中的檔案:
檔案 大小格式 
U0001-2308202217491500.pdf6.74 MBAdobe PDF檢視/開啟
顯示文件簡單紀錄


系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。

社群連結
聯絡資訊
10617臺北市大安區羅斯福路四段1號
No.1 Sec.4, Roosevelt Rd., Taipei, Taiwan, R.O.C. 106
Tel: (02)33662353
Email: ntuetds@ntu.edu.tw
意見箱
相關連結
館藏目錄
國內圖書館整合查詢 MetaCat
臺大學術典藏 NTU Scholars
臺大圖書館數位典藏館
本站聲明
© NTU Library All Rights Reserved