Skip navigation

DSpace JSPUI

DSpace preserves and enables easy and open access to all types of digital content including text, images, moving images, mpegs and data sets

Learn More
DSpace logo
English
中文
  • Browse
    • Communities
      & Collections
    • Publication Year
    • Author
    • Title
    • Subject
    • Advisor
  • Search TDR
  • Rights Q&A
    • My Page
    • Receive email
      updates
    • Edit Profile
  1. NTU Theses and Dissertations Repository
  2. 電機資訊學院
  3. 電信工程學研究所
Please use this identifier to cite or link to this item: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/88569
Full metadata record
???org.dspace.app.webui.jsptag.ItemTag.dcfield???ValueLanguage
dc.contributor.advisor王鈺強zh_TW
dc.contributor.advisorYu-Chiang Frank Wangen
dc.contributor.author紀彥仰zh_TW
dc.contributor.authorYan-Yang Jien
dc.date.accessioned2023-08-15T16:52:40Z-
dc.date.available2023-11-09-
dc.date.copyright2023-08-15-
dc.date.issued2023-
dc.date.submitted2023-07-28-
dc.identifier.citationS. Zakharov, W. Kehl, A. Bhargava, and A. Gaidon, “Autolabeling 3d objects with differentiable rendering of sdf shape priors,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020. vi, 1, 2, 6, 7, 17, 18, 19, 20
J. Leonard, J. How, S. Teller, M. Berger, S. Campbell, G. Fiore, L. Fletcher, E. Frazzoli, A. Huang, S. Karaman et al., “A perception-driven autonomous urban vehicle,” Journal of Field Robotics, 2008. 1
H. Gao, B. Cheng, J. Wang, K. Li, J. Zhao, and D. Li, “Object classification using cnn-based fusion of vision and lidar in autonomous vehicle environment,” IEEE Transactions on Industrial Informatics, 2018. 1
W. Lee, N. Park, and W. Woo, “Depth-assisted real-time 3d object detection for augmented reality,” in Proceedings of the International Conference on Artificial Reality and Telexistence, 2011. 1
S. Shi, X. Wang, and H. Li, “Pointrcnn: 3d object proposal generation and detection from point cloud,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019. 1, 4, 17, 20
A. H. Lang, S. Vora, H. Caesar, L. Zhou, J. Yang, and O. Beijbom, “Pointpillars: Fast encoders for object detection from point clouds,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019. 1, 4, 17
X. Ma, S. Liu, Z. Xia, H. Zhang, X. Zeng, and W. Ouyang, “Rethinking pseudo-lidar representation,” in Proceedings of the European Conference on Computer Vision (ECCV), 2020. 1, 4, 17
S. Song, S. P. Lichtenberg, and J. Xiao, “Sun rgb-d: A rgb-d scene understanding benchmark suite,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015. 1
H. Wang, Y. Cong, O. Litany, Y. Gao, and L. J. Guibas, “3dioumatch: Leveraging iou prediction for semi-supervised 3d object detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021. 1, 5
J. Wang, H. Gang, S. Ancha, Y.-T. Chen, and D. Held, “Semi-supervised 3d object detection via temporal graph neural networks,” in International Conference on 3D Vision (3DV), 2021. 1, 5
J. Yin, J. Fang, D. Zhou, L. Zhang, C.-Z. Xu, J. Shen, and W. Wang, “Semi-supervised 3d object detection with proficient teachers,” in Proceedings of the European Conference on Computer Vision (ECCV), 2022. 1, 5
Z. Qin, J. Wang, and Y. Lu, “Weakly supervised 3d object detection from point clouds,” in Proceedings of the ACM Conference on Multimedia (MM), 2020. 1, 2, 6, 7, 17, 18
Q. Meng, W. Wang, T. Zhou, J. Shen, L. Van Gool, and D. Dai, “Weakly supervised 3d object detection from lidar point cloud,” in Proceedings of the European Conference on Computer Vision (ECCV), 2020. 1, 5
Y. Wei, S. Su, J. Lu, and J. Zhou, “Fgr: Frustum-aware geometric reasoning for weakly supervised 3d vehicle detection,” in Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 2021. 1, 6, 7
R. McCraith, E. Insafutdinov, L. Neumann, and A. Vedaldi, “Lifting 2d object locations to 3d by discounting lidar outliers across objects and views,” in Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 2022. 1, 2, 6, 7, 10, 17, 18
L. Peng, S. Yan, B. Wu, Z. Yang, X. He, and D. Cai, “Weakm3d: Towards weakly supervised monocular 3d object detection,” in Proceedings of the International Conference on Learning Representations (ICLR), 2022. 1, 2, 6, 7, 17, 18
H. Wang, S. Sridhar, J. Huang, J. Valentin, S. Song, and L. J. Guibas, “Normalized object coordinate space for category-level 6d object pose and size estimation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019. 2, 6
J. J. Park, P. Florence, J. Straub, R. Newcombe, and S. Lovegrove, “Deepsdf: Learning continuous signed distance functions for shape representation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019. 2, 6
M. A. Fischler and R. C. Bolles, “Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography,” Communications of the ACM, 1981. 6
F. Lu, Z. Liu, X. Song, D. Zhou, W. Li, H. Miao, M. Liao, L. Zhang, B. Zhou, R. Yang et al., “Permo: Perceiving more at once from a single image for autonomous driving,” arXiv preprint arXiv:2007.08116, 2020. 8
S. Wold, K. Esbensen, and P. Geladi, “Principal component analysis,” Chemometrics and intelligent laboratory systems, 1987. 8
J. Zbontar, L. Jing, I. Misra, Y. LeCun, and S. Deny, “Barlow twins: Self-supervised learning via redundancy reduction,” in Proceedings of the International Conference on Machine Learning (ICML), 2021. 14, 22
A. Geiger, P. Lenz, and R. Urtasun, “Are we ready for autonomous driving? the kitti vision benchmark suite,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2012. 16
X. Chen, H. Ma, J. Wan, B. Li, and T. Xia, “Multi-view 3d object detection network for autonomous driving,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017. 16
A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga et al., “Pytorch: An imperative style, high-performance deep learning library,” 2019. 16
C. R. Qi, H. Su, K. Mo, and L. J. Guibas, “Pointnet: Deep learning on point sets for 3d classification and segmentation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017. 16
D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014. 17
T. Chen, S. Kornblith, M. Norouzi, and G. Hinton, “A simple framework for contrastive learning of visual representations,” in Proceedings of the International Conference on Machine Learning (ICML), 2020. 21, 22
-
dc.identifier.urihttp://tdr.lib.ntu.edu.tw/jspui/handle/123456789/88569-
dc.description.abstract三維物體偵測是三維視覺的一個熱門研究領域,近年來受到廣泛關注。然而,訓練用於三維物體偵測的深度學習模型通常需要大量帶有三維邊界框註釋的數據,這是一項耗時的任務並且存在重大挑戰。為了應對這一挑戰,我們提出了一種通過可變形模板匹配(DTMNet)進行弱監督三維物體偵測的方法,該方法在圖像和二維實例遮罩的弱監督下,通過將可變形形狀模板與輸入的LiDAR點雲進行匹配,生成弱監督的三維虛擬邊界框。生成的三維虛擬邊界框可以用於訓練基於圖像或基於LiDAR的三維物體偵測器。我們的DTMNet顯著降低了註釋成本,提高了三維物體偵測的效率。對KITTI基準數據集的實驗結果在定量和定性上證明了我們提出的模型的有效性和實用性zh_TW
dc.description.abstract3D object detection is an active research topic for 3D vision and has been widely studied in recent years. However, training deep learning models for 3D object detection typically requires extensive data with 3D bounding box annotations, which is a time-consuming task and presents a significant challenge. To address this challenge, we propose a weakly supervised 3D object detection method via deformable template matching (DTMNet), which generates weakly supervised 3D pseudo-bounding boxes by matching a deformable shape template with the input LiDAR point clouds under the weak supervision of images and 2D instance masks. The generated 3D pseudo-bounding boxes can be used to train either image-based or LiDAR-based 3D object detectors. Our DTMNet significantly reduces annotation costs and improves the efficiency of 3D object detection. Experimental results on the KITTI benchmark dataset quantitatively and qualitatively demonstrate the effectiveness and practicality of our proposed model.en
dc.description.provenanceSubmitted by admin ntu (admin@lib.ntu.edu.tw) on 2023-08-15T16:52:40Z
No. of bitstreams: 0
en
dc.description.provenanceMade available in DSpace on 2023-08-15T16:52:40Z (GMT). No. of bitstreams: 0en
dc.description.tableofcontents中文摘要 i
Abstract ii
List of Figures v
List of Tables vii
1 Introduction 1
2 Related Work 4
2.1 Supervised 3D object detection 4
2.2 Semi-supervised 3D object detection 5
2.3 Weakly supervised 3D object detection 6
3 Proposed Method 8
3.1 Problem formulation and model overview 8
3.2 Weakly supervised deformable template matching 10
3.2.1 Segmentor and Predictor 10
3.2.2 Edge and color supervision 12
3.3 Training and obtaining pseudo-bounding box 15
4 Experiments 16
4.1 Dataset and implementation details 16
4.1.1 Dataset 16
4.1.2 Implementation Details 16
4.2 Weakly supervised 3D object detection 17
4.2.1 Quantitative evaluation 17
4.2.2 Qualitative result 18
4.3 Ablation Study 20
4.4 Additional Experiment Results 22
5 Conclusion 25
Reference 26
-
dc.language.isoen-
dc.subject物體偵測zh_TW
dc.subject三維視覺zh_TW
dc.subject點雲zh_TW
dc.subjectPoint Clouden
dc.subjectObject Detectionen
dc.subject3D Visionen
dc.title基於可變形模板匹配之弱監督三維物體檢測zh_TW
dc.titleWeakly Supervised 3D Object Detection via Deformable Template Matchingen
dc.typeThesis-
dc.date.schoolyear111-2-
dc.description.degree碩士-
dc.contributor.oralexamcommittee孫紹華;陳祝嵩zh_TW
dc.contributor.oralexamcommitteeShao-Hua Sun;Chu-Song Chenen
dc.subject.keyword物體偵測,三維視覺,點雲,zh_TW
dc.subject.keywordObject Detection,3D Vision,Point Cloud,en
dc.relation.page29-
dc.identifier.doi10.6342/NTU202302077-
dc.rights.note同意授權(全球公開)-
dc.date.accepted2023-08-01-
dc.contributor.author-college電機資訊學院-
dc.contributor.author-dept電信工程學研究所-
Appears in Collections:電信工程學研究所

Files in This Item:
File SizeFormat 
ntu-111-2.pdf10.23 MBAdobe PDFView/Open
Show simple item record


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

社群連結
聯絡資訊
10617臺北市大安區羅斯福路四段1號
No.1 Sec.4, Roosevelt Rd., Taipei, Taiwan, R.O.C. 106
Tel: (02)33662353
Email: ntuetds@ntu.edu.tw
意見箱
相關連結
館藏目錄
國內圖書館整合查詢 MetaCat
臺大學術典藏 NTU Scholars
臺大圖書館數位典藏館
本站聲明
© NTU Library All Rights Reserved