請用此 Handle URI 來引用此文件:
http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/21293
標題: | 基於焦點移轉的方式理解指稱表達式中的物件關係 Referring Relationships Comprehension by Residual Attention Shift |
作者: | Ying-Wei Wu 吳楹偉 |
指導教授: | 徐宏民 |
關鍵字: | 指稱關係,物件偵測,深度學習, Referring relationship,Object detection,Deep learning, |
出版年 : | 2019 |
學位: | 碩士 |
摘要: | 物件偵測這個問題在電腦視覺領域裏已經被研究很久了,也有許多重
大的成果。近年來人們開始把重心放在理解影像中的物件關係上,這樣子 的問題叫做物件關係偵測,在這幾年大家的努力下已經有一些成果。SSAS [2] 提出了我們應該要把物件關係偵測這個問題倒過來做,也 就是給定物件關係再去偵測物件,原因是這樣的問題更符合人類直覺,他們 把這個問題稱作理解指稱表達式中的物件關係。我們認為 SSAS [2] 提出的 想法確實合理,並且能夠輕易地應用在人類和機器人的互動上。在這篇研究 中我們探討了 SSAS [2] 這篇論文中的缺陷,並且提出改進方法,我們還根 據自己的觀察,設計出一個全新的網路元件。在我們儘可能公平的比較下, 我們是在這個問題上表現最好的方法,也確實解決了 SSAS [2] 上的諸多問 題。 Object detection is an already well researched topic in the area of com- puter vision. Recently, people tried to pay their attention to the visual relationships between objects, the task is called visual relationship detection, there were plenty works that made progress in these years. SSAS [2] claimed that we should focus on the inverse problem of visual relationship detection, which is to detect the object from a given relationship. They thought the inverse problem is more conform to human intuition, and called this referring relationship. We thought that their statements were reasonable, and it can be applied to human robot interaction scenario easily. In this thesis, we explored the drawback of SSAS [2], and proposed our solution, we also designed a new network component with our observation. With a comparison as fair as we can make, we achieved state-of-the-art performance in the task, and indeed resolved the problems of SSAS [2]. |
URI: | http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/21293 |
DOI: | 10.6342/NTU201903405 |
全文授權: | 未授權 |
顯示於系所單位: | 資訊工程學系 |
文件中的檔案:
檔案 | 大小 | 格式 | |
---|---|---|---|
ntu-108-1.pdf 目前未授權公開取用 | 5.13 MB | Adobe PDF |
系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。