Skip navigation

DSpace

機構典藏 DSpace 系統致力於保存各式數位資料(如:文字、圖片、PDF)並使其易於取用。

點此認識 DSpace
DSpace logo
English
中文
  • 瀏覽論文
    • 校院系所
    • 出版年
    • 作者
    • 標題
    • 關鍵字
    • 指導教授
  • 搜尋 TDR
  • 授權 Q&A
    • 我的頁面
    • 接受 E-mail 通知
    • 編輯個人資料
  1. NTU Theses and Dissertations Repository
  2. 電機資訊學院
  3. 電機工程學系
請用此 Handle URI 來引用此文件: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/68398
完整後設資料紀錄
DC 欄位值語言
dc.contributor.advisor傅立成
dc.contributor.authorShih-Hsi Hsuen
dc.contributor.author徐世曦zh_TW
dc.date.accessioned2021-06-17T02:19:48Z-
dc.date.available2020-09-04
dc.date.copyright2017-09-04
dc.date.issued2017
dc.date.submitted2017-08-21
dc.identifier.citation[1] Bonin-Font, Francisco, Alberto Ortiz, and Gabriel Oliver. “Visual navigation for Mobile robots: A survey.” Journal of intelligent and robotic systems 53.3 (2008): 263.
[2] Sim, Robert, and Gregory Dudek. “Learning generative models of scene features.” International Journal of Computer Vision 60.1 (2004): 45-61.
[3] Sim, Robert, and Gregory Dudek. 'Effective exploration strategies for the construction of visual maps.' Intelligent Robots and Systems, (2003).
[4] Sim, Robert, Pantelis Elinas, Matt Griffin, Alex Shyr and James J. Little. “Design and analysis of a framework for real-time vision-based SLAM using Rao-Blackwellised particle filters.” Computer and Robot Vision, (2006).
[5] Sim, Robert, and James J. Little. “Autonomous vision-based exploration and mapping using hybrid maps and Rao-Blackwellised particle filters.” Intelligent Robots and Systems, (2006).
[6] De Cristóforis, Pablo, Matias Nitsche, Tomáš Krajník, Taihú Pire, and Marta Mejail. “Hybrid vision-based navigation for mobile robots in mixed indoor/outdoor environments.” Pattern Recognition Letters 53 (2015): 118-128.
[7] Labbe, Mathieu, and François Michaud. “Online global loop closure detection for large-scale multi-session graph-based slam.” Intelligent Robots and Systems (2014).
[8] Gartshore, R., A. Aguado, and C. Galambos. “Incremental map building using an occupancy grid for an autonomous monocular robot.” Control, Automation, Robotics and Vision, (2002).
[9] Gartshore, Rachel, Phil Palmer, and John Illingworth. “A novel exploration algorithm based on a improvement strategy.” International Journal of Advanced Robotic Systems 2.4 (2005): 30.
[10] Talukder, Ashit, and Larry Matthies. “Real-time detection of moving objects from moving vehicles using dense stereo and optical flow.” Intelligent Robots and Systems, (2004).
[11] Srinivasan, Mandyam V. Saul Thurrowgood, and Dean Soccol. “An optical system for guidance of terrain following in UAVs.” Video and Signal Based Surveillance, (2006).
[12] Michels, Jeff, Ashutosh Saxena, and Andrew Y. Ng. “High speed obstacle avoidance using monocular vision and reinforcement learning.” Proceedings of the 22nd international conference on Machine learning. ACM, (2005).
[13] Cherubini, Andrea, Fabien Spindler, and François Chaumette. “Autonomous visual navigation and laser-based moving obstacle avoidance.” IEEE Transactions on Intelligent Transportation Systems 15.5 (2014): 2101-2110.
[14] Wang, Tianmiao, Yicheng Zhang, Chaolei Wang, Jianhong Liang, Han Gao, Miao Liu, Qinpu Guan, and Anqi Sun. “Indoor visual navigation system based on paired-landmark for small UAVs.” Robotics and Biomimetics, (2014).
[15] Cheng, Chen, Wennan Chai, and Hubert Roth. “A single frame depth visual gyroscope and its integration for robot navigation and mapping in structured indoor environments.” Journal of Intelligent & Robotic Systems 80.3-4 (2015): 365-374.
[16] Lei, Tai, Shaohua Li, and Ming Liu. “A deep-network solution towards model-less obstacle avoidance.” Intelligent Robots and Systems, (2016).
[17] Yang, Shichao, Sandeep Konam, Chen, Ma, Stephanie, Rosenthal, Manuela, Veloso, and Sebastian, Scherer. “Obstacle Avoidance through Deep Networks based Intermediate Perception.” arXiv preprint arXiv:1704.08759 (2017).
[18] Gupta, Saurabh, James Davidson, Sergey Levine, Rahul Sukthankar, and Jitendra Malik. “Cognitive mapping and planning for visual navigation.” arXiv preprint arXiv:1702.03920 (2017).
[19] Oh, Junhyuk, Valliappa Chockalingam, Satinder Singh, and Honglak Lee. “Control of memory, active perception, and action in minecraft.” International Conference on Machine Learning. (2016).
[20] Mirowski, Piotr, Razvan Pascanu, Fabio Viola, Hubert Soyer, Andrew J. Ballard, Andrea Banino, Misha Denil, Ross Goroshin, Laurent Sifre, Koray Kavukcuoglu, Dharshan Kumaran, and Raia Hadsel. “Learning to navigate in complex environments.” arXiv preprint arXiv:1611.03673 (2016).
[21] Zhu, Yuke, Roozbeh Mottaghi, Eric Kolve, Joseph J. Lim, Abhinav Gupta, Li Fei-Fei, and Ali, Farhadi. “Target-driven visual navigation in indoor scenes using deep reinforcement learning.” Robotics and Automation (ICRA), (2017).
[22] Zhang, Jingwei, Jost T. Springenberg, Joschka Boedecker, and Wolfram Burgard. “Deep Reinforcement Learning with Successor Features for Navigation across Similar Environments.” arXiv preprint arXiv:1612.05533 (2016).
[23] Sutton, Richard S., and Andrew G. Barto. “Introduction to reinforcement learning.” Vol. 135. Cambridge: MIT Press, (1998).
[24] Watkins, CJCH. “Learning from delayed rewards.” King's College, Cambridge, (1989).
[25] Watkins, CJCH., and Peter Dayan. “Q-learning.” Machine learning 8.3-4 (1992): 279-292.
[26] http://www.cc.ntu.edu.tw/chinese/epaper/0038/20160920_3805.html
[27] https://en.wikipedia.org/wiki/Sigmoid_function
[28] https://zh.wikipedia.org/wiki/%E7%BA%BF%E6%80%A7%E6%95%B4%E6%B5%81%E5%87%BD%E6%95%B0
[29] https://deeplearning4j.org/neuralnet-overview#custom
[30] http://ithelp.ithome.com.tw/articles/10187424
[31] LeCun, Yann, Léon Bottou, Yoshua Bengio, and Patrick Haffner. “Gradient-based learning applied to document recognition.” Proceedings of the IEEE 86.11 (1998): 2278-2324.
[32] Alex, Krizhevsky, Ilya Sutskever, and Geoffrey, E. Hinton. “Imagenet classification with deep convolutional neural networks.” Advances in neural information processing systems. (2012).
[33] Simonyan, Karen, and Andrew Zisserman. “Very deep convolutional networks for large-scale image recognition.” arXiv preprint arXiv:1409.1556 (2014).
[34] Szegedy, Christian, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew Rabinovich. “Going deeper with convolutions.” Proceedings of the IEEE conference on computer vision and pattern recognition, (2015).
[35] Mnih, Volodymyr, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan Wierstra, and Martin Riedmiller. “Playing atari with deep reinforcement learning.” arXiv preprint arXiv:1312.5602 (2013).
[36] Mnih, Volodymyr, Koray Kavukcuoglu, David Silver, Andrei A. Rusu, Joel Veness, Marc G. Bellemare, Alex Graves,Martin Riedmiller, Andreas K. Fidjeland, Georg Ostrovski, Stig Petersen, Charles Beattie, Amir Sadik, Ioannis Antonoglou, Helen King, Dharshan Kumaran, Daan Wierstra, Shane Legg, and Demis Hassabis. “Human-level control through deep reinforcement learning.” Nature 518.7540 (2015): 529-533.
[37] Mnih, Volodymyr, Adrià P. Badia, Mehdi Mirza, Alex Graves, Tim Harley, Timothy P. Lillicrap, David Silver, and Koray Kavukcuoglu. “Asynchronous methods for deep reinforcement learning.” International Conference on Machine Learning. (2016).
[38] https://medium.com/emergent-future/simple-reinforcement-learning-with-tensorflow-part-8-asynchronous-actor-critic-agents-a3c-c88f72a5e9f2
[39] Jaderberg, Max, Volodymyr Mnih, Wojciech M. Czarnecki, Tom Schaul, Joel Z. Leibo, David Silver, and Koray Kavukcuoglu. “Reinforcement learning with unsupervised auxiliary tasks.” arXiv preprint arXiv:1611.05397 (2016).
[40] Stuart, Lloyd. “Least squares quantization in PCM.” IEEE transactions on information theory 28.2 (1982): 129-137.
[41] Masci, Jonathan, Ueli Meier, Dan Cireşan, and Jürgen Schmidhuber. “Stacked convolutional auto-encoders for hierarchical feature extraction.” Artificial Neural Networks and Machine Learning–ICANN 2011 (2011): 52-59.
[42] Armeni, Iro, Ozan Sener,Amir R. Zamir, Helen Jiang, Ioannis Brilakis, Martin Fischer, and Silvio Savarese. “3d semantic parsing of large-scale indoor spaces.” Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. (2016).
[43] https://matterport.com
[44] http://doc.aldebaran.com/2-4/home_pepper.html
dc.identifier.urihttp://tdr.lib.ntu.edu.tw/jspui/handle/123456789/68398-
dc.description.abstract在傳統的視覺室內導航中,問題常常被敘述為如何從影像感測轉換成控制策略。而主要研究以分成兩種,一種方法會利用影像和預先建好的地圖來進行室內導航,此種方法可以達成全域的最佳規劃,然而地圖建置費時又需要大量的運算,而當環境變動時也不容易即時的更新;另一種方法則是嘗試只利用即時影像的方式來控制移動,可是此方法往往受限於資訊的局部性,而無法達成全域的最佳規劃。近年來深度學習相關的研究興起,許多領域受益於此而得以獲得長足的進步,而最近又以深度強化學習最受矚目。在室內導航問題中,深度強化學習可以幫助機器人將複雜的環境影像轉換成馬達控制的指令,藉以到達目的地。
本論文提出了一個新穎的架構,其目標在於達成無須地圖資訊也可以讓機器人學習如何在寬廣的室內空間中導航。寬廣空間中的室內導航往往牽涉到複雜的空間理解,尤其是室內空間往往有諸多的牆壁,門,遮蔽住視線,使得影像輸入往往複雜且難以分析。借助分散式深度強化學習與自動編碼器所產生的圖像嵌入空間,本方法可以實現機器人在寬廣的空間導航,不需藉助額外的地圖與人類的指導,可以到達指定的目的地。在實驗的部分中我們用模擬的室內環境和現實的室內環境檢驗了提出的方法,並且成功的達成了導航的任務。
zh_TW
dc.description.abstractIn the traditional vision based indoor navigation domain, the problem is usually described as how to convert the visual perception to control policy. The main research can be categorized into two types, one of which is the method using pre-constructed map and current perception to accomplish indoor visual navigation. This type of methods can easily achieve global optimal path planning, however, mapping algorithm is usually time-consuming and needs a lot of computation power. Also, this type of methods are sensitive to the changes in the environment and the map cannot be easily updated in a short time. Another types of method are navigation based on only real time image, yet this type of methods is restricted to local perception, and it is hard to achieve global optimal planning. Recently, as the rising of deep learning related researches, many research fields make a progress, and the deep reinforcement learning gets the most attention. In the indoor visual navigation problem, deep reinforcement learning can help the robot to convert the complicated environment scene to motor control command, and accomplish the navigation task.
In this thesis, we propose a novel structure, where objective is to achieve large-scale environment navigation in the indoor environment without pre-constructed map. The large-scale indoor environment needs good understanding to work for complex spatial perception, especially when the indoor space consists of many walls and doors which might occlude the view of robot. By the proposed distributed deep reinforcement learning and image embedding space generated by auto-encoder, out method can achieve large-scale indoor visual navigation without extra map information and human instruction. In the experiments, we show the validation of our proposed method by conducting successful navigation tasks both in simulation and real environments.
en
dc.description.provenanceMade available in DSpace on 2021-06-17T02:19:48Z (GMT). No. of bitstreams: 1
ntu-106-R04921010-1.pdf: 2881193 bytes, checksum: 98ea43df4f6e66fdb3919c2267de6e15 (MD5)
Previous issue date: 2017
en
dc.description.tableofcontentsCONTENTS
口試委員審定書 i
誌謝 ii
摘要 iii
Abstract iv
contents vi
List of Figures viii
List of Tables x
Chapter 1 Introduction 1
1.1 Motivation 1
1.2 Related Works 3
1.2.1 Indoor Visual Navigation 3
1.2.2 Visual Navigation with Deep Learning 7
1.3 Challenges 9
1.4 Objectives 10
1.5 Thesis Organization 11
Chapter 2 Preliminary 12
2.1 Reinforcement Learning 12
2.1.1 Value-function method 13
2.1.2 Q-Learning 16
2.2 Deep Learning 17
2.2.1 Deep Neural Network 17
2.2.2 Convolutional Neural Network 20
2.3 Deep Reinforcement Learning 21
2.3.1 Deep Q Network 22
2.3.2 Asynchronous advantage actor and critic algorithm 23
Chapter 3 Distributed Map-less Visual Navigation based on DRL 26
3.1 Problem Formulation 26
3.2 System Architecture 27
3.3 Auto-encoder and Auxiliary Task 31
3.4 Deep Reinforcement Learning model 33
3.5 Distributed Map-less Visual Navigation 37
Chapter 4 Experiments 40
4.1 Training in Simulation 40
4.1.1 Simulation Environment 40
4.1.2 Learning Setup 44
4.2 Evaluation in Simulation 46
4.3 Training in Real Environment 53
4.4 Evaluation in Real Environment 55
Chapter 5 Conclusion 57
Reference 59
dc.language.isoen
dc.subject圖像嵌入zh_TW
dc.subject深度強化學習zh_TW
dc.subject無地圖室內視覺導航zh_TW
dc.subjectDeep reinforcement learningen
dc.subjectimage embeddingen
dc.subjectmap-less indoor visual navigationen
dc.title基於圖像嵌入與深度強化學習之無地圖室內視覺導航zh_TW
dc.titleMap-less Indoor Visual Navigation based on Image Embedding and Deep Reinforcement Learningen
dc.typeThesis
dc.date.schoolyear105-2
dc.description.degree碩士
dc.contributor.oralexamcommittee周瑞仁,簡忠漢,林其禹,范欽雄
dc.subject.keyword深度強化學習,圖像嵌入,無地圖室內視覺導航,zh_TW
dc.subject.keywordDeep reinforcement learning,image embedding,map-less indoor visual navigation,en
dc.relation.page62
dc.identifier.doi10.6342/NTU201704104
dc.rights.note有償授權
dc.date.accepted2017-08-21
dc.contributor.author-college電機資訊學院zh_TW
dc.contributor.author-dept電機工程學研究所zh_TW
顯示於系所單位:電機工程學系

文件中的檔案:
檔案 大小格式 
ntu-106-1.pdf
  未授權公開取用
2.81 MBAdobe PDF
顯示文件簡單紀錄


系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。

社群連結
聯絡資訊
10617臺北市大安區羅斯福路四段1號
No.1 Sec.4, Roosevelt Rd., Taipei, Taiwan, R.O.C. 106
Tel: (02)33662353
Email: ntuetds@ntu.edu.tw
意見箱
相關連結
館藏目錄
國內圖書館整合查詢 MetaCat
臺大學術典藏 NTU Scholars
臺大圖書館數位典藏館
本站聲明
© NTU Library All Rights Reserved