Skip navigation

DSpace JSPUI

DSpace preserves and enables easy and open access to all types of digital content including text, images, moving images, mpegs and data sets

Learn More
DSpace logo
English
中文
  • Browse
    • Communities
      & Collections
    • Publication Year
    • Author
    • Title
    • Subject
    • Advisor
  • Search TDR
  • Rights Q&A
    • My Page
    • Receive email
      updates
    • Edit Profile
  1. NTU Theses and Dissertations Repository
  2. 電機資訊學院
  3. 電信工程學研究所
Please use this identifier to cite or link to this item: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/92674
Title: 增強 3D 場景理解的通用語意神經輻射場
GSNeRF: Generalizable Semantic Neural Radiance Fields with Enhanced 3D Scene Understanding
Authors: 周子庭
Zi-Ting Chou
Advisor: 王鈺強
Yu-Chiang Frank Wang
Keyword: 深度學習,三維電腦視覺,神經輻射場,語意分割,
Deep Learning,3D Computer Vision,Neural Radiance Field,Semantic Segmentation,
Publication Year : 2024
Degree: 碩士
Abstract: 利用多視角輸入合成新視角圖像,神經輻射場(Neural Radiance Fields,簡稱NeRF)已成為三維視覺領域中的熱門研究課題。在本文中,我們引入了一種名為通用語義神經輻射場(Generalizable Semantic Neural Radiance Fields,簡稱GSNeRF)的方法,它在合成過程中獨特地考慮了圖像語義,從而可以為未見過的場景生成新視角圖像及其相關的語義地圖。我們的GSNeRF由兩個階段組成:語義幾何推理和深度引導的視覺渲染。前者能夠觀察多視角圖像輸入,從場景中提取語義和幾何特徵。後者在圖像幾何信息的指導下,執行圖像和語義渲染,具有更好的性能。我們的實驗不僅證實了GSNeRF在新視角圖像和語義分割合成方面優於先前的工作,而且進一步驗證了我們的採樣策略對視覺渲染的有效性。
Utilizing multi-view inputs to synthesize novel-view images, Neural Radiance Fields (NeRF) have emerged as a popular research topic in 3D vision. In this work, we introduce a Generalizable Semantic Neural Radiance Fields (GSNeRF), which uniquely takes image semantics into the synthesis process so that both novel view image and the associated semantic maps can be produced for unseen scenes. Our GSNeRF is composed of two stages: Semantic Geo-Reasoning and Depth-Guided Visual rendering. The former is able to observe multi-view image inputs to extract semantic and geometry features from a scene. Guided by the resulting image geometry information, the latter performs both image and semantic rendering with improved performances. Our experiments not only confirm that GSNeRF performs favorably against prior works on both novel-view image and semantic segmentation synthesis but the effectiveness of our sampling strategy for visual rendering is further verified.
URI: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/92674
DOI: 10.6342/NTU202400922
Fulltext Rights: 同意授權(限校園內公開)
Appears in Collections:電信工程學研究所

Files in This Item:
File SizeFormat 
ntu-112-2.pdf
Access limited in NTU ip range
1.22 MBAdobe PDF
Show full item record


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

社群連結
聯絡資訊
10617臺北市大安區羅斯福路四段1號
No.1 Sec.4, Roosevelt Rd., Taipei, Taiwan, R.O.C. 106
Tel: (02)33662353
Email: ntuetds@ntu.edu.tw
意見箱
相關連結
館藏目錄
國內圖書館整合查詢 MetaCat
臺大學術典藏 NTU Scholars
臺大圖書館數位典藏館
本站聲明
© NTU Library All Rights Reserved