請用此 Handle URI 來引用此文件:
http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/86996完整後設資料紀錄
| DC 欄位 | 值 | 語言 |
|---|---|---|
| dc.contributor.advisor | 周俊廷 | zh_TW |
| dc.contributor.advisor | Chun-Ting Chou | en |
| dc.contributor.author | 黃冠仁 | zh_TW |
| dc.contributor.author | Kuan-Jen Huang | en |
| dc.date.accessioned | 2023-05-02T17:18:53Z | - |
| dc.date.available | 2023-11-09 | - |
| dc.date.copyright | 2023-05-02 | - |
| dc.date.issued | 2023 | - |
| dc.date.submitted | 2023-01-13 | - |
| dc.identifier.citation | Thalen, J.P. (2006). ADAS for the Car of the Future (Bachelor's thesis, University of Twente).
Howard, I. P., & Rogers, B. J. (1995). Binocular vision and stereopsis. Oxford University Press, USA. Koller, D., Luong, Q. T., & Malik, J. (1994, October). Using binocular stereopsis for vision-based vehicle control. In Proceedings of the Intelligent Vehicles' 94 Symposium (pp. 237-242). IEEE. Uttamchandani, D. (Ed.). (2013). Handbook of MEMS for wireless and mobile applications. Elsevier. “What is lidar?” https://velodynelidar.com/what-is-lidar/ Khader, M., & Cherian, S. (2020). An introduction to automotive lidar. Texas Instruments. Atapour-Abarghouei, A., & Breckon, T. P. (2019, September). Monocular segment-wise depth: Monocular depth estimation based on a semantic segmentation prior. In 2019 IEEE International Conference on Image Processing (ICIP) (pp. 4295-4299). IEEE. Lee, J. H., Han, M. K., Ko, D. W., & Suh, I. H. (2019). From big to small: Multi-scale local planar guidance for monocular depth estimation. arXiv preprint arXiv:1907.10326v6. Garg, R., Bg, V. K., Carneiro, G., & Reid, I. (2016, October). Unsupervised cnn for single view depth estimation: Geometry to the rescue. In European conference on computer vision (pp. 740-756). Springer, Cham. Godard, C., Mac Aodha, O., & Brostow, G. J. (2017). Unsupervised monocular depth estimation with left-right consistency. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 270-279). Stein, G. P., Mano, O., & Shashua, A. (2003, June). Vision-based ACC with a single camera: bounds on range and range rate accuracy. In IEEE IV2003 intelligent vehicles symposium. Proceedings (Cat. No. 03TH8683) (pp. 120-125). IEEE. Gat, I., Benady, M., & Shashua, A. (2005). A monocular vision advance warning system for the automotive aftermarket. SAE transactions, 403-410. Qi, S. H., Li, J., Sun, Z. P., Zhang, J. T., & Sun, Y. (2019, February). Distance estimation of monocular based on vehicle pose information. In Journal of Physics: Conference Series (Vol. 1168, No. 3, p. 032040). IOP Publishing. Bao, C., Chen, C., Kui, H., & Wang, X. (2019, June). Safe driving at traffic lights: An image recognition based approach. In 2019 20th IEEE International Conference on Mobile Data Management (MDM) (pp. 112-117). IEEE. “道路交通標誌標線號誌設置規則, 第158條” https://law.moj.gov.tw/LawClass/LawSingle.aspx?pcode=K0040014&flno=158 Zhang, Z. (1999, September). Flexible camera calibration by viewing a plane from unknown orientations. In Proceedings of the seventh IEEE international conference on computer vision (Vol. 1, pp. 666-673). IEEE. Ajsmilutin. (2017). CarND-Advanced-Lane-Lines. Github. https://github.com/ajsmilutin/CarND-Advanced-Lane-Lines Camera Calibration and 3D Reconstruction. https://docs.opencv.org/4.x/d9/d0c/group__calib3d.html#ga69f2545a8b62a6b0fc2ee060dc30559d | - |
| dc.identifier.uri | http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/86996 | - |
| dc.description.abstract | 距離預測技術在未來智慧城市中扮演重要的角色,透過預測物體的距離,可以發展出各式各樣的應用,包含自動駕駛系統、車輛定位和街景圖資的更新等,在這些應用中,距離預測都是不可或缺的技術。在現有的距離預測技術中,有一類方法採用感測器如無線電雷達(Radar)或是光學雷達(Lidar),藉由對周圍物體發出光或電磁波來預測距離,這類方法雖然精準且快但是成本較為昂貴。另種方法則是使用影像等低成本的方式,再經由強大的演算法來預測出影像中物體的距離。
在本篇論文中,我們嘗試透過單一鏡頭相機(Monocular Camera)來估計物體距離。此類問題可被細分為兩個種類包含地面上物體和相機的距離,以及非地面物體和相機的距離,像是預測紅綠燈的距離。大多數現有的研究專注在預測地面上物體的距離且這些研究的結果都很準確與穩定。而第二類的問題通常需要額外的資訊才能被解決,因此我們提出使用目標物體的真實高度與相機成像模型(Camera Imaging Model)進而推導出物體在影像中的座標與此物體對應之地面點座標的關係。透過這個關係,我們可以推出任何已知高度的物體之地面座標,進而利用現有處理第一類問題的演算法來預測此地面物體與相機的距離。 透過本文所提出的演算法,不但可以用低成本且即時的方式來預測出單一影像中已知高度的物體之距離,此方法更具有資料獨立性,也就是效能不受使用的資料不同而影響。實驗結果顯示我們的方法在三個不同環境中的整體平均絕對百分比誤差(Mean Absolute Percentage Error)為8%,在估計非地面物體之距離比基於學習(Learning-based)的模型好30%。但是我們的方法無法使用在預測低高度物體上,且需要一個額外的物件偵測模型來獲得物體在影像中的位置。 | zh_TW |
| dc.description.abstract | Distance estimation technology is an essential component of future smart cities. It can be used to enable a variety of applications such as autonomous driving systems, and map updating. There are various methods for distance estimation, including using active sensors like radar and lidar, which emit waves or light to measure the distance to surrounding objects. These methods are generally accurate and fast, but they can also be expensive. Alternatively, low-cost techniques such as using images can be combined with powerful algorithms to estimate the distance of objects in the image.
In this thesis, we try to estimate the distance of objects through a monocular camera. This problem can be divided into two categories. One is estimation of distance between the objects on the ground and the camera. The other is the estimation of distance between the objects above the ground and the camera, such as traffic lights. Most of the existing researches focus on the former and the estimation is quite accurate and stable. Solving the latter problem usually requires additional information such as the object size. Our method utilizes the height of the target object and camera imaging model to derive the relationship between the coordinate of an object in the image plane and its projection point on the ground. By this relationship, we can calculate the coordinates of the object’s projection point on the ground. Then, the existing algorithm that is used in the first type of problems is adopted to estimate the distance of the projection point. By using our proposed method, the distance of an object with a known height in the monocular image can be predicted in a low-cost and real-time way. Moreover, our method is data-independent. The experimental results show that the overall mean absolute percentage error of our proposed method in three different environments is 8%, which outperforms the learning-based model by 30%. However, our method cannot be used to estimate the distance of objects with a low height and needs an additional object detection model to obtain the position of an object in the image. | en |
| dc.description.provenance | Submitted by admin ntu (admin@lib.ntu.edu.tw) on 2023-05-02T17:18:53Z No. of bitstreams: 0 | en |
| dc.description.provenance | Made available in DSpace on 2023-05-02T17:18:53Z (GMT). No. of bitstreams: 0 | en |
| dc.description.tableofcontents | 誌謝 i
中文摘要 ii ABSTRACT iii CONTENTS v CHAPTER 1 INTRODUCTION 1 1.1 Motivation 1 1.2 Related Work 5 1.2.1 Sensor-based Method 5 1.2.2 Learning-based Method 7 1.2.3 Geometry-based Method 8 1.2.4 Summary of the Related Work 11 1.3 Problem Statement 12 1.4 Contributions 13 1.5 Thesis Organization 13 CHAPTER 2 SYSTEM SETTINGS AND ASSUMPTIONS 14 2.1 Input Data of Driving video 14 2.2 Key Ideas of the Proposed Method 15 CHAPTER 3 PROPOSED ALGORITHM 17 3.1 Pipeline of the Proposed Solutions 17 3.2 Step 1: Data Preprocessing - Camera Calibration 19 3.3 Step 2: Estimate the Ground Point of Any Object with a Known Heigh23 3.3.1 Step 2.1: Coordinate from Real World to Image Plane 23 3.3.2 Step 2.2: The Difference of Y-coordinate in the image 25 3.3.3 Step 2.3: Relationship between Object Position and Difference of Y-coordinate 27 3.3.4 Step 2.4: Estimate the Ground Point of Any Object with a Known Height 38 3.4 Step 3: Perspective Transform 43 3.4.1 Step 3.1: Define the Region of the Bird-eye Image 43 3.4.2 Step 3.2: Generate the Bird-eye Image 47 3.4.2.1 Get Homography Matrix 47 3.4.2.2 Warping the Image into a Bird-eye Image 49 3.4.3 Step 3.3: Pixel Resolution of Ground 51 3.5 Step 4: Distance Estimation 55 CHAPTER 4 PERFORMANCE EVALUATION 57 4.1 Ground Truth Datasets & Settings 57 4.1.1 Description of Ground Truth Dataset 57 4.1.2 Performance Metrics 69 4.2 Experimental Results 71 4.2.1 Performance of Estimating the Distance of Objects in Dataset_1 71 4.2.1.1 Predict the Length of White Dashed Lines on the Road 72 4.2.1.2 Predict the Distance of Objects on the Indoor Ground 73 4.2.2 Performance of Estimating the Distance of Objects in Dataset_2 75 4.3 Estimating the Distance of Objects without a Precise Height 79 CHAPTER 5 CONCLUSIONS 86 REFERENCES 87 | - |
| dc.language.iso | en | - |
| dc.subject | 相機幾何 | zh_TW |
| dc.subject | 單一影像距離預測 | zh_TW |
| dc.subject | 視角轉換 | zh_TW |
| dc.subject | perspective transform | en |
| dc.subject | monocular distance estimation | en |
| dc.subject | camera geometry | en |
| dc.title | 使用單目相機對已知高度物體深度預測 | zh_TW |
| dc.title | Depth Estimation of Objects with Known Heights using a Monocular Camera | en |
| dc.type | Thesis | - |
| dc.date.schoolyear | 111-1 | - |
| dc.description.degree | 碩士 | - |
| dc.contributor.oralexamcommittee | 逄愛君;魏宏宇;莊永裕 | zh_TW |
| dc.contributor.oralexamcommittee | Ai-Chun Pang;Hung-Yu Wei;Yung-Yu Chuang | en |
| dc.subject.keyword | 單一影像距離預測,相機幾何,視角轉換, | zh_TW |
| dc.subject.keyword | monocular distance estimation,camera geometry,perspective transform, | en |
| dc.relation.page | 88 | - |
| dc.identifier.doi | 10.6342/NTU202300097 | - |
| dc.rights.note | 同意授權(全球公開) | - |
| dc.date.accepted | 2023-01-14 | - |
| dc.contributor.author-college | 電機資訊學院 | - |
| dc.contributor.author-dept | 電信工程學研究所 | - |
| 顯示於系所單位: | 電信工程學研究所 | |
文件中的檔案:
| 檔案 | 大小 | 格式 | |
|---|---|---|---|
| ntu-111-1.pdf | 35.39 MB | Adobe PDF | 檢視/開啟 |
系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。
