請用此 Handle URI 來引用此文件:
http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/87481完整後設資料紀錄
| DC 欄位 | 值 | 語言 |
|---|---|---|
| dc.contributor.advisor | 劉浩澧 | zh_TW |
| dc.contributor.advisor | Hao-Li Liu | en |
| dc.contributor.author | 羅敏宏 | zh_TW |
| dc.contributor.author | Min-Hung Lo | en |
| dc.date.accessioned | 2023-06-13T16:12:45Z | - |
| dc.date.available | 2026-02-03 | - |
| dc.date.copyright | 2023-06-13 | - |
| dc.date.issued | 2023 | - |
| dc.date.submitted | 2023-02-06 | - |
| dc.identifier.citation | [1] M. Najafi, P. Abolmaesumi, and R. Rohling, “Single-camera closed-form real-time needle tracking for ultrasound-guided needle insertion,” Ultrasound in Medicine Biology, vol. 41, no. 10, pp. 2663–2676, 2015. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0301562915003658
[2] H. Hiroto, “Digging into detectron 2,” https://medium.com/@hirotoschwert/digging-into-detectron-2-47b2e794fabd. [3] O. Güler and Z. Yaniv, “Image-guided navigation: a cost effective practical introduction using the image-guided surgery toolkit (igstk),” Conference proceedings :Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Conference, vol. 2012, pp. 6056–9, 08 2012. [4] E. Basafa, M. Hoßbach, and P. Stolka, “Fast, intuitive, vision-based: Performance metrics for visual registration, instrument guidance, and image fusion,” 10 2016, pp.9–17. [5] J. C. Gadsden, M. Latmore, and D. M. Levine, “Evaluation of the ezono 4000 with ezguide for ultrasound-guided procedures,” Expert Review of Medical Devices, vol. 12, pp. 251 – 261, 2015. [6] A. Hakime, A. Barah, F. Deschamps, G. Farouil, J. Joskin, L. Tselikas, A. Auperin,and T. de Baere, “Prospective comparison of freehand and electromagnetic needle tracking for us-guided percutaneous liver biopsy,” Journal of Vascular and Interventional Radiology, vol. 24, no. 11, pp. 1682–1689, 2013. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S1051044313010282 [7] G. Bradski, “The OpenCV Library,” Dr. Dobb’s Journal of Software Tools, 2000. [8] Z.-Y. Chiu, A. Z. Liao, F. Richter, B. Johnson, and M. C. Yip, “Markerless suture needle 6d pose tracking with robust uncertainty estimation for autonomous minimally invasive robotic surgery,” ArXiv, vol. abs/2109.12722, 2021. [9] F. Salehahmadi and F. Hajialiasgari, “Grand adventure of augmented reality in landscape of surgery,” World journal of plastic surgery, vol. 8, pp. 135–145, 05 2019. [10] A. Elmi-Terander, H. Skulason, M. Söderman, J. Racadio, R. Homan, D. Babic, N. van der Vaart, and R. Nachabe, “Surgical navigation technology based on augmented reality and integrated 3d intraoperative imaging: A spine cadaveric feasibility and accuracy study.” Spine, Nov. 2016. [11] C. R. Harris, K. J. Millman, S. J. van der Walt, R. Gommers, P. Virtanen, D. Cournapeau, E. Wieser, J. Taylor, S. Berg, N. J. Smith, R. Kern, M. Picus, S. Hoyer, M. H. van Kerkwijk, M. Brett, A. Haldane, J. F. del Río, M. Wiebe, P. Peterson, P. Gérard-Marchant, K. Sheppard, T. Reddy, W. Weckesser, H. Abbasi, C. Gohlke, and T. E. Oliphant, “Array programming with NumPy, Nature, vol. 585, no. 7825, pp. 357–362, Sep. 2020. [Online]. Available: https://doi.org/10.1038/s41586-020-2649-2 [12] Y. Abdel-Aziz, H. Karara, and M. Hauck, “Direct linear transformation from comparator coordinates into object space coordinates in close-range photogrammetry*,”Photogrammetric Engineering Remote Sensing, vol. 81, no. 2, pp.103–107, 2015. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0099111215303086 [13] X.-S. Gao, X.-R. Hou, J. Tang, and H.-F. Cheng, “Complete solution classification for the perspective-three-point problem,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 25, no. 8, pp. 930–943, 2003. [14] V. Lepetit, F. Moreno-Noguer, and P. Fua, “Epnp: An accurate o(n) solution to the pnp problem,” International Journal of Computer Vision, vol. 81, pp. 155–166, 2008. [15] T. Lin, M. Maire, S. J. Belongie, L. D. Bourdev, R. B. Girshick, J. Hays, P. Perona, D. Ramanan, P. Doll’a r, and C. L. Zitnick, “Microsoft COCO: common objects in context,” CoRR, vol. abs/1405.0312, 2014. [Online]. Available: http://arxiv.org/abs/1405.0312 [16] M. Andriluka, L. Pishchulin, P. Gehler, and B. Schiele, “2d human pose estimation: New benchmark and state of the art analysis,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2014. [17] D. DeTone, T. Malisiewicz, and A. Rabinovich, “Superpoint: Self-supervised interest point detection and description,” 2017, cite arxiv:1712.07629 Comment: Camera-ready version for CVPR 2018 Deep Learning for Visual SLAM Workshop(DL4VSLAM2018). [Online]. Available: http://arxiv.org/abs/1712.07629 [18] Y. Wu, A. Kirillov, F. Massa, W.-Y. Lo, and R. Girshick, “Detectron2,” https://github.com/facebookresearch/detectron2, 2019. [19] J. E. Bresenham, “Algorithm for computer control of a digital plotter,” IBM Systems Journal, vol. 4, no. 1, pp. 25–30, 1965. [20] P. T. Boggs and J. R. Donaldson, “Orthogonal distance regression,” 1989. [21] G. An, S. Lee, M.-W. Seo, K. Yun, W.-S. Cheong, and S.-J. Kang, “Charuco boardbased omnidirectional camera calibration method,” Electronics, vol. 7, p. 421, 12 2018. | - |
| dc.identifier.uri | http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/87481 | - |
| dc.description.abstract | 針具穿刺手術是目前醫療界中相當常見的手術方式,例如射頻燒灼術(RFA) 與活體組織檢查(biopsy)。然而,要讓針具精準地穿刺到指定細胞位置並遵循特定軌跡,醫師們必須在手術過程中採用實時醫療影像或相機影像作為導引,例如超音波影像與針具軌跡預測。然而在針具穿刺過程中,針具軌跡無法被精準預測,這可能會導致手術過程不順暢,甚至給病人帶來危害。其中一大阻礙因素就是針具只能在有限的相機視野內追蹤。因此,我們提出了一種混合性的標記與針具姿態估測擴增實境系統,透過同時追蹤整個針具組合,提升手術的可靠性,並擴展可操作的工作空間。這個系統不僅能夠透過標記追蹤計算手把姿態,還能透過針體的特徵點來估算空間資訊,擴大可偵測的範圍。我們所提出的系統實現了3.13 倍的空間偵測範圍擴展,整體準確度為2.9±0.01mm,此外系統還能夠提供針具彎折的警示,避免器材偏誤造成的手術風險。我們提出的方法使用CCD相機來實現基於AR的手術導引,可以為目前的RFA手術帶來增益效果。 | zh_TW |
| dc.description.abstract | Surgical needle insertion is a commonly performed medical procedure, such as radiofrequency ablation (RFA) and needle biopsy. To ensure that the needle is accurately inserted into the targeted tissue following a designated trajectory, clinicians often rely on real-time image feedback for needle trajectory visualization and prediction using ultrasonography and/or optical camera feedback. However, during needle puncture guiding procedures, the needle trajectory may not always be accurately reported, which can cause challenges or even hazards to the system. One of the major obstacles is that the needle can only be tracked within a limited camera field of view. In this work, we proposed a hybrid pose estimation augmented reality (AR) algorithm that allows for simultaneous tracking of the entire needle set, thus expanding the detectable range of the needle insertion procedure. The algorithm not only tracks the 3D pose of the needle handle (using marker tracking) but also tracks the pose of the needle itself (using key point detection), thereby increasing the trackable zone during the procedure. The detectable view is measured to have up to 3.13 times improvement in the field-of-view, with an overall accuracy of 2.9±0.01 mm. This proposed algorithm represents an advancement in using CCD camera as a tool for implementing AR-based surgical guidance and may be beneficial for current surgical procedures such as RFA or biopsy. | en |
| dc.description.provenance | Submitted by admin ntu (admin@lib.ntu.edu.tw) on 2023-06-13T16:12:44Z No. of bitstreams: 0 | en |
| dc.description.provenance | Made available in DSpace on 2023-06-13T16:12:45Z (GMT). No. of bitstreams: 0 | en |
| dc.description.tableofcontents | 口試委員審定書i
致謝i 摘要ii Abstract iii 圖目錄vii 表目錄ix 第一章緒論1 1.1 影像導引手術系統. . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1.2 肝癌微創手術: 射頻燒灼術(Radiofrequency Ablation) . . . . . . . . 2 1.3 針具穿刺視覺化導引之相關研究. . . . . . . . . . . . . . . . . . . . 3 1.4 研究目的與貢獻. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 第二章方法與理論7 2.1 系統總覽. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 2.1.1 系統架構. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 2.2 系統校正. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 2.2.1 相機校正. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 2.2.2 針具校正. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 2.3 混合型姿態估測. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 2.3.1 基於解析幾何的姿態估測. . . . . . . . . . . . . . . . . . . . . . 12 2.3.2 基於標記的姿態估測. . . . . . . . . . . . . . . . . . . . . . . . 12 2.3.3 基於共線特徵點的針具姿態估測. . . . . . . . . . . . . . . . . . 13 2.4 針具特徵點偵測與優化. . . . . . . . . . . . . . . . . . . . . . . . . 15 2.4.1 深度學習用於特徵點偵測. . . . . . . . . . . . . . . . . . . . . . 15 2.4.2 特徵點優化. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20 2.5 針具彎曲檢測. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23 2.6 擴增實境視覺化. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 2.7 評估指標. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27 第三章實驗設置與結果31 3.1 實驗目的. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 3.2 實驗設置. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 3.3 實驗結果. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32 3.3.1 前處理校正結果. . . . . . . . . . . . . . . . . . . . . . . . . . . 32 3.3.2 特徵點偵測誤差與優化結果. . . . . . . . . . . . . . . . . . . . 32 3.3.3 姿態估測誤差. . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 3.3.4 手術視野擴展度. . . . . . . . . . . . . . . . . . . . . . . . . . . 37 3.3.5 針具演算法效益. . . . . . . . . . . . . . . . . . . . . . . . . . . 38 3.3.6 偵測率. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39 3.3.7 針具彎曲指標. . . . . . . . . . . . . . . . . . . . . . . . . . . . 40 3.3.8 實時處理實現(Real-time implementation) . . . . . . . . . . . . . 43 3.3.9 成效比較. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43 第四章結論與未來展望44 4.1 結論. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44 4.2 未來展望. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45 參考文獻. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .46 附錄A — 實作演示影片. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .50 | - |
| dc.language.iso | zh_TW | - |
| dc.subject | 針具穿刺手術導引 | zh_TW |
| dc.subject | 姿態估測 | zh_TW |
| dc.subject | 擴增實境 | zh_TW |
| dc.subject | 標記追蹤 | zh_TW |
| dc.subject | 特徵點檢測 | zh_TW |
| dc.subject | Augmented reality | en |
| dc.subject | Pose estimation | en |
| dc.subject | Marker-based tracking | en |
| dc.subject | Surgical needle insertion guidance | en |
| dc.subject | Key points detection | en |
| dc.title | 混合型姿態估測於針具穿刺手術導引的擴增實境系統 | zh_TW |
| dc.title | Hybrid Pose Estimation Augmented Reality System in Assisting Surgical Needle Insertion Guidance | en |
| dc.type | Thesis | - |
| dc.date.schoolyear | 111-1 | - |
| dc.description.degree | 碩士 | - |
| dc.contributor.oralexamcommittee | 邱錫彥;龍震宇;李正匡 | zh_TW |
| dc.contributor.oralexamcommittee | Shin-Yan Chiou;Chen-Yu Lung;Cheng-Kuang Lee | en |
| dc.subject.keyword | 針具穿刺手術導引,擴增實境,姿態估測,標記追蹤,特徵點檢測, | zh_TW |
| dc.subject.keyword | Surgical needle insertion guidance,Augmented reality,Pose estimation,Marker-based tracking,Key points detection, | en |
| dc.relation.page | 50 | - |
| dc.identifier.doi | 10.6342/NTU202300160 | - |
| dc.rights.note | 同意授權(限校園內公開) | - |
| dc.date.accepted | 2023-02-08 | - |
| dc.contributor.author-college | 電機資訊學院 | - |
| dc.contributor.author-dept | 電機工程學系 | - |
| dc.date.embargo-lift | 2026-02-03 | - |
| 顯示於系所單位: | 電機工程學系 | |
文件中的檔案:
| 檔案 | 大小 | 格式 | |
|---|---|---|---|
| ntu-111-1.pdf 授權僅限NTU校內IP使用(校園外請利用VPN校外連線服務) | 13.9 MB | Adobe PDF |
系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。
